ORIGINAL RESEARCH article
Front. Comput. Sci.
Sec. Networks and Communications
Optimizing Cloud Data Center Performance with Demand-Responsive Load-Balancing Scheme
Provisionally accepted- The Federal Polytechnic Ilaro, Ilaro, Nigeria
Select one of your emails
You have multiple emails registered with Frontiers:
Notify me on publication
Please enter your email address:
If you already have an account, please login
You don't have a Frontiers account ? You can register here
Cloud computing has revolutionized modern computing by providing scalable, on-demand access to computing resources through a pay-as-you-use model, accessible from any location. As businesses increasingly migrate their operations to cloud platforms, cloud service providers continue to expand data center infrastructures to support a growing number of concurrent users. However, one persistent challenge identified in literature is the issue of uneven load distribution, which often leads to performance degradation, increased latency, and reduced resource availability. While several load balancing techniques have been proposed to address this issue, many still suffer from inefficiencies, such as delayed redirection of loads from overloaded to underutilized servers, and the inability to meet key load balancing metrics, including scalability, makespan, support for heterogeneous environments, and fault tolerance. To address these limitations, this study proposes a demand-responsive dynamic load balancing scheme that optimizes cloud data center performance by incorporating real-time server metrics and modeling user request arrivals using Poisson distribution. The scheme dynamically assigns incoming loads based on server capacity, current workload, and estimated request arrival rates, thereby reducing server overload and minimizing response delays. The simulation was implemented using load-balancing data extracted from the GoCJ (Google Cloud Jobs) dataset and evaluated using Python. Comparative performance analysis with existing techniques such as Round Robin, Least Connection, and static threshold-based methods demonstrated the superiority of the proposed scheme. Quantitative results revealed that the proposed method achieved an average response time of 10–19 ms, compared to 18–30 ms for Round Robin and 15–25 ms for Least Connection algorithms. Furthermore, the proposed scheme reduced the number of failed or delayed user requests by over 40%, improved server utilization balance, and maintained robust performance under high-load scenarios. These findings confirm the effectiveness of the proposed demand-based load balancing strategy in enhancing the responsiveness, scalability, and efficiency of cloud data centers, thereby supporting the growing demands of cloud users and applications
Keywords: Assignment problem, Cloud computing, data center, Game theory, Loadbalancing, Queue theory
Received: 22 Sep 2024; Accepted: 26 Dec 2025.
Copyright: © 2025 Soyemi and Hammed. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
* Correspondence: Jumoke Soyemi
Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.