Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Comput. Sci.

Sec. Networks and Communications

Optimizing Cloud Data Center Performance with Demand-Responsive Load-Balancing Scheme

Provisionally accepted
  • The Federal Polytechnic Ilaro, Ilaro, Nigeria

The final, formatted version of the article will be published soon.

Cloud computing has revolutionized modern computing by providing scalable, on-demand access to computing resources through a pay-as-you-use model, accessible from any location. As businesses increasingly migrate their operations to cloud platforms, cloud service providers continue to expand data center infrastructures to support a growing number of concurrent users. However, one persistent challenge identified in literature is the issue of uneven load distribution, which often leads to performance degradation, increased latency, and reduced resource availability. While several load balancing techniques have been proposed to address this issue, many still suffer from inefficiencies, such as delayed redirection of loads from overloaded to underutilized servers, and the inability to meet key load balancing metrics, including scalability, makespan, support for heterogeneous environments, and fault tolerance. To address these limitations, this study proposes a demand-responsive dynamic load balancing scheme that optimizes cloud data center performance by incorporating real-time server metrics and modeling user request arrivals using Poisson distribution. The scheme dynamically assigns incoming loads based on server capacity, current workload, and estimated request arrival rates, thereby reducing server overload and minimizing response delays. The simulation was implemented using load-balancing data extracted from the GoCJ (Google Cloud Jobs) dataset and evaluated using Python. Comparative performance analysis with existing techniques such as Round Robin, Least Connection, and static threshold-based methods demonstrated the superiority of the proposed scheme. Quantitative results revealed that the proposed method achieved an average response time of 10–19 ms, compared to 18–30 ms for Round Robin and 15–25 ms for Least Connection algorithms. Furthermore, the proposed scheme reduced the number of failed or delayed user requests by over 40%, improved server utilization balance, and maintained robust performance under high-load scenarios. These findings confirm the effectiveness of the proposed demand-based load balancing strategy in enhancing the responsiveness, scalability, and efficiency of cloud data centers, thereby supporting the growing demands of cloud users and applications

Keywords: Assignment problem, Cloud computing, data center, Game theory, Loadbalancing, Queue theory

Received: 22 Sep 2024; Accepted: 26 Dec 2025.

Copyright: © 2025 Soyemi and Hammed. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

* Correspondence: Jumoke Soyemi

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.