Open Access

Queuing-Integrated Deep Reinforcement Learning For Adaptive Task Scheduling In Cloud Data Centers

4 Technical University of Munich, Germany

Abstract

The accelerating digitalization of economic, industrial, and social systems has rendered cloud computing the backbone of contemporary information infrastructure. Yet, the unprecedented growth in computational demand, heterogeneity of workloads, and volatility of user requirements have exposed deep limitations in classical task scheduling and resource management paradigms. Static or heuristics-based schedulers, which historically dominated cloud environments, are increasingly unable to cope with highly dynamic and stochastic workloads, fluctuating service-level requirements, and the imperative for energy-efficient operations. This study advances a comprehensive theoretical and analytical investigation of deep reinforcement learning–driven dynamic task scheduling in cloud computing, with particular emphasis on queuing-aware optimal decision making. Building on the methodological foundation established by Kanikanti et al. (2025), who demonstrated the effectiveness of deep Q-learning combined with optimal queuing theory for cloud task scheduling, this research situates their contribution within a broader interdisciplinary framework that spans energy-aware systems, multi-agent learning, and cyber-physical digital twins.

The article develops a unifying perspective that integrates insights from reinforcement learning theory, stochastic queuing models, energy management in cyber-physical systems, and adaptive control of autonomous agents. By synthesizing developments in microgrid energy management, underwater robotics, autonomous vehicle control, and digital twin–based production systems, the study demonstrates that the core challenge of cloud scheduling is not merely computational efficiency but the orchestration of learning-driven decisions across uncertain, delayed, and resource-constrained environments. In this sense, cloud data centers resemble complex adaptive systems in which computing tasks compete for shared resources in a manner analogous to energy flows in microgrids or coordinated actions in multi-robot systems.

Methodologically, the research adopts a text-based analytical design that combines formal reinforcement learning principles derived from Markov decision processes with queuing-theoretic interpretations of cloud workloads. The deep Q-learning framework of Kanikanti et al. (2025) is critically analyzed and extended conceptually through comparative evaluation against SARSA-based, actor–critic, and deep deterministic policy gradient approaches reported in the broader literature. Particular attention is devoted to how state abstraction, reward shaping, and queue length feedback enable schedulers to balance latency, throughput, and energy consumption simultaneously.

The results of this study are presented in a descriptive and interpretive manner grounded in the comparative literature. They indicate that deep Q-learning–based dynamic schedulers consistently outperform rule-based and shallow reinforcement learning approaches in terms of adaptive responsiveness, queue stability, and energy-aware decision making, as supported by studies in cloud computing, microgrids, and robotic coordination. The discussion further reveals that queuing-informed deep reinforcement learning architectures provide a theoretically robust mechanism for mitigating congestion collapse, improving quality of service, and aligning cloud operations with sustainability goals.

By offering an extensive theoretical elaboration and critical synthesis of existing research, this article contributes a unified conceptual framework for understanding and advancing learning-driven cloud task scheduling. It concludes that the convergence of deep reinforcement learning and optimal queuing theory, as exemplified by Kanikanti et al. (2025), represents not a marginal technical improvement but a paradigm shift in how future cloud ecosystems will be designed, governed, and optimized.

Keywords

References

📄 Ariche, S.; Boulghasoul, Z.; Ouardi, A.E.; Elbacha, A.; Tajer, A.; Espie, S. Enhancing Energy Management in Battery Electric Vehicles: A Novel Approach Based on Fuzzy Q-Learning Controller. Eng. Sci. Technol. Int. J. 2025, 67, 102070.
📄 Yu, T.; Huang, J.; Chang, Q. Optimizing task scheduling in human-robot collaboration with deep multiagent reinforcement learning. Journal of Manufacturing Systems 2021, 60, 487–499.
📄 Liang, Z.B.; Li, Q.; Fu, G.D. Multi-UAV Collaborative Search and Attack Mission Decision-Making in Unknown Environments. Sensors 2023, 23, 7398.
📄 Kanikanti, V. S. N.; Tiwari, S. K.; Nayan, V.; Suryawanshi, S.; Chauhan, R. Deep Q-Learning Driven Dynamic Optimal Task Scheduling for Cloud Computing Using Optimal Queuing. In 2025 International Conference on Computational Intelligence and Knowledge Economy, 2025, 217–222.
📄 Ramesh, S.; Sukanth, B.N.; Sathyavarapu, S.J.; Sharma, V.; Kumaar, A.A.N.; Khanna, M. Comparative Analysis of Q-Learning, SARSA, and Deep Q-Network for Microgrid Energy Management. Sci. Rep. 2025, 15, 694.
📄 Ding, D.; Fan, X.; Zhao, Y.; Kang, K.; Yin, Q.; Zeng, J. Q-learning based dynamic task scheduling for energy-efficient cloud computing. Future Generation Computer Systems 2020.
📄 Puterman, M.L. Markov Decision Processes: Discrete Stochastic Dynamic Programming; John Wiley and Sons: Hoboken, NJ, USA, 1994.
📄 Sutton, R.; Barto, A. Reinforcement Learning: An Introduction; MIT Press: Cambridge, MA, USA, 1998.
📄 Guo, W.; Tian, W.; Ye, Y.; Xu, L.; Wu, K. Cloud Resource Scheduling with Deep Reinforcement Learning and Imitation Learning. IEEE Internet of Things Journal 2020.
📄 Carlucho, I.; De Paula, M.; Wang, S.; Petillot, Y.; Acosta, G.G. Adaptive Low-Level Control of Autonomous Underwater Vehicles Using Deep Reinforcement Learning. Robot. Auton. Syst. 2018, 107, 71–86.
📄 Luo, Y.; Ball, P. Adaptive Production Strategy in Vertical Farm Digital Twins with Q-Learning Algorithms. Sci. Rep. 2025, 15, 15129.
📄 Xu, Y.J.; Li, H. Secondary Voltage Control Strategy for DC Microgrid Based on Reinforcement Learning. Mech. Electr. Eng. Technol. 2025, 54, 173–178.
📄 Mohan Sharma; Garg, R. An artificial neural network based approach for energy efficient task scheduling in cloud data centers. Elsevier 2020.
📄 Rybak, L.A.; Behera, L.; Averbukh, M.A.; Sapryka, A.V. Development of an Algorithm for Managing a Multi-Robot System for Cargo Transportation Based on Reinforcement Learning in a Virtual Environment. IOP Conf. Ser. Mater. Sci. Eng. 2020, 945, 012083.
📄 Korivand, S.; Galvani, G.; Ajoudani, A.; Gong, J.; Jalili, N. Optimizing Human–Robot Teaming Performance through Q-Learning-Based Task Load Adjustment and Physiological Data Analysis. Sensors 2024, 24, 2817.
📄 Sufan, V.; Troni, G. Swim4Real: Deep Reinforcement Learning-Based Energy-Efficient and Agile 6-DOF Control for Underwater Vehicles. IEEE Robot. Autom. Lett. 2025, 10, 7326–7333.
📄 Asghari, A.; Sohrabi, M.K.; Yaghmaee, F. Task scheduling, resource provisioning, and load balancing on scientific workflows using parallel SARSA reinforcement learning agents and genetic algorithm. Journal of Supercomputing 2020.
📄 Chen, D.; Wang, H.; Hu, D.; Xian, Q.; Wu, B. Q-Learning Improved Golden Jackal Optimization Algorithm and Its Application to Reliability Optimization of Hydraulic System. Sci. Rep. 2024, 14, 24587.
📄 Dong, X.; Zhang, H.; Xie, X.; Ming, Z. Data-Driven Distributed H Infinity Current Sharing Consensus Optimal Control of DC Microgrids via Reinforcement Learning. IEEE Trans. Circuits Syst. Regul. Pap. 2024, 71, 2824–2834.
📄 Li, X.X. Research on Energy Management Strategies for Fuel Cell Hybrid Electric Ships. Wuhan University of Technology, Wuhan, China, 2023.
📄 Wang, J.J.; Zhou, H.M.; Guo, J.; Si, H.W.; Xu, C.; Zhang, M.H.; Zhang, Y.Q.; Zhou, G.X. A Q-Learning-Based Deep Deterministic Policy Gradient Algorithm for the Re-Entrant Hybrid Flow Shop Joint Scheduling Problem with Dual-Gripper. Eng. Lett. 2025, 33, 1632–1647.
📄 Wang, X.; Zhu, Q.X.; Zhu, Y.H.; Miao, L.Y. Path Planning for Mobile Robots Based on Improved Q-Learning Algorithm. Comput. Simul. 2025, 42, 371–377.
📄 Mostafavi, S.; Hakami, V. A Stochastic Approximation Approach for Foresighted Task Scheduling in Cloud Computing. Wireless Personal Communications 2020.
📄 Qi, Q.; Zhang, L.; Wang, J.; Sun, H.; Zhuang, Z.; Liao, J.; Yu, F.R. Scalable Parallel Task Scheduling for Autonomous Driving Using Multi-Task Deep Reinforcement Learning. IEEE Transactions on Vehicular Technology 2020.
📄 Zhu, H.; Li, M.; Tang, Y.; Sun, Y. A Deep Reinforcement-Learning-Based Optimization Approach for Real-Time Scheduling in Cloud Manufacturing. IEEE Access 2020.

Similar Articles

1-10 of 19

You may also start an advanced similarity search for this article.