Articles | Open Access |

Queuing-Aware Deep Reinforcement Learning For Intelligent Task Scheduling In Large-Scale Cloud Computing Systems

John Miller , Department of Computer Science, University of Toronto, Canada

Abstract

Cloud computing has evolved into a foundational paradigm for modern digital infrastructures, supporting an ever-expanding range of computational workloads across enterprise, scientific, and consumer domains. The exponential growth of cloud services, coupled with highly dynamic and heterogeneous user demands, has rendered traditional static and heuristic-based scheduling mechanisms increasingly insufficient. Contemporary cloud systems are now required to deliver not only high performance but also economic efficiency, quality of service compliance, and energy-aware operation under volatile workload conditions. Within this context, the integration of intelligent learning-based methods with classical queuing theory has emerged as a promising research frontier. This article develops a comprehensive and theoretically grounded examination of adaptive task scheduling in cloud computing through the synthesis of deep reinforcement learning and optimal queuing models.

The study is motivated by the recent emergence of deep Q-learning driven scheduling frameworks that explicitly incorporate queuing dynamics into the learning process, as exemplified by Kanikanti et al. (2025), who demonstrate how deep Q-learning combined with optimal queuing can significantly improve dynamic task allocation in cloud environments. While earlier cloud scheduling research relied on heuristic, evolutionary, or analytically derived models, such as those discussed by Tawfeek et al. (2013), Beloglazov and Buyya (2012), and Vilaplana et al. (2014), these approaches often struggle to adapt to highly non-stationary and uncertain workloads. Reinforcement learning, by contrast, provides a formal framework for agents to learn optimal control policies through interaction with the environment, enabling continuous adaptation to changing system states. When reinforced by queuing theory, which offers mathematically grounded representations of waiting times, congestion, and service capacity, deep Q-learning becomes a powerful mechanism for capturing both short-term dynamics and long-term system objectives.

This research article develops a unified conceptual and methodological framework that situates deep Q-learning-based scheduling within the broader theoretical landscape of cloud computing, queuing systems, and resource optimization. Drawing upon the foundational perspectives of Armbrust et al. (2010) on cloud service models and economic drivers, as well as the rigorous queuing formulations of Kleinrock (1975) and Khazaei et al. (2012), the paper establishes a rich theoretical basis for modeling cloud data centers as learning-enabled service systems. The methodology articulates how cloud workloads, virtual machines, and service queues can be encoded into a state-action-reward structure suitable for deep Q-learning, while also acknowledging the constraints and assumptions that shape such models. Rather than presenting numerical simulations or experimental tables, the study provides a detailed interpretive analysis grounded in the existing literature, exploring how intelligent schedulers informed by queuing theory can outperform conventional heuristics in terms of responsiveness, stability, and cost efficiency.

The results section offers a descriptive synthesis of how deep Q-learning-based optimal queuing approaches are expected to influence task completion times, resource utilization, and service level adherence, drawing heavily on prior analytical and empirical insights from hybrid cloud scheduling, auto-scaling, and optimization research. The discussion then situates these findings within a broader scholarly debate, examining the epistemological shift from rule-based to learning-based scheduling, the potential risks of model instability and training overhead, and the future implications for hybrid and multi-cloud ecosystems. Ultimately, this article argues that the convergence of deep reinforcement learning and queuing theory represents not merely a technical enhancement but a paradigmatic transformation in how cloud infrastructures are designed, managed, and optimized in the face of accelerating complexity.

Keywords

Cloud computing, task scheduling, deep Q-learning

References

Armbrust, M., Fox, A., Griffith, R., Joseph, A. D., Katz, R., Konwinski, A., and Zaharia, M. A view of cloud computing. Communications of the ACM, 53(4), 50-58.

Kanikanti, V. S. N., Tiwari, S. K., Nayan, V., Suryawanshi, S., and Chauhan, R. Deep Q-Learning Driven Dynamic Optimal Task Scheduling for Cloud Computing Using Optimal Queuing. In 2025 International Conference on Computational Intelligence and Knowledge Economy (ICCIKE), IEEE.

Sindhu, S., and Mukherjee, S. Efficient task scheduling algorithms for cloud computing environment. In High Performance Architecture and Grid Computing, Springer.

Tawfeek, M. A., El-Sisi, A., Keshk, A. E., and Torkey, F. A. Cloud task scheduling based on ant colony optimization. In International Conference on Computer Engineering and Systems.

Sharieh, A., and Al-Thwaib, E. A mathematical model for hybrid multi cloud environment. In International Conference on Information Technology.

Beloglazov, A., and Buyya, R. Optimal online deterministic algorithms and adaptive heuristics for energy and performance efficient dynamic consolidation of virtual machines in cloud data centers. Concurrency and Computation Practice and Experience, 24(13), 1397-1420.

Mao, M., Li, J., and Humphrey, M. Cloud auto-scaling with deadline and budget constraints. In IEEE ACM International Conference on Grid Computing.

Vilaplana, J., Solsona, F., Teixido, I., Mateo, J., Abella, F., and Rius, J. A queuing theory model for cloud computing. Journal of Supercomputing, 68(1), 492-507.

Kleinrock, L. Queueing systems theory. Wiley Interscience.

Goyal, S. Public vs private vs hybrid vs community cloud computing. International Journal of Computer Network and Information Security, 6(3), 20.

Islam, M. S. U., and Rana, B. Task scheduling in cloud computing. International Journal of Advance Research Ideas and Innovations in Technology, 3(4), 642-646.

Van den Bossche, R., Vanmechelen, K., and Broeckhove, J. Cost optimal scheduling in hybrid iaas clouds for deadline constrained workloads. In IEEE International Conference on Cloud Computing.

Khazaei, H., Misic, J., and Misic, V. B. Performance analysis of cloud computing centers using M G m m r queuing systems. IEEE Transactions on Parallel and Distributed Systems, 23(5), 936-943.

Agarwal, D., and Jain, S. Efficient optimal algorithm of task scheduling in cloud computing environment. arXiv preprint.

Lakra, A. V., and Yadav, D. K. Multiobjective tasks scheduling algorithm for cloud computing throughput optimization. Procedia Computer Science, 48, 107-113.

Branch, K. A novel task scheduling method in cloud environment using cuckoo optimization algorithm. International Journal of Cloud Computing and Super Computing, 2(2), 7-20.

Kaur, N., and Chhabra, A. Comparative analysis of job scheduling algorithms in parallel and distributed computing environments. International Journal of Advanced Research in Computer Science, 8(3).

Sharieh, A., and Albdour, L. A heuristic approach for service allocation in cloud computing. International Journal of Cloud Applications and Computing, 7(4), 60-74.

Grandinetti, L. Pervasive cloud computing technologies future outlooks and interdisciplinary perspectives. IGI Global.

Gabriel Herrera. SaaS 101 IaaS PaaS SaaS and cloud computing. StartHQ Blog.

Vijay, P., and Anju, B. G. An efficient workflow scheduling approach in cloud computing. Doctoral dissertation.

Al-Shaikh, A., Khattab, H., Sharieh, A., and Sleit, A. Resource utilization in cloud computing as an optimization problem. International Journal of Advanced Computer Science and Applications, 7(6), 336-342.

Kaleeswaran, A., Ramasamy, V., and Vivekanandan, P. Dynamic scheduling of data using genetic algorithm in cloud computing. Park College of Engineering and Technology.

Gogulan, R., Kavitha, A., and Kumar, U. K. A multiple pheromone algorithm for cloud scheduling with various quality of service requirements. International Journal of Computer Science.

Article Statistics

Downloads

Download data is not yet available.

Copyright License

Download Citations

How to Cite

John Miller. (2026). Queuing-Aware Deep Reinforcement Learning For Intelligent Task Scheduling In Large-Scale Cloud Computing Systems. International Journal of Modern Medicine, 5(01), 46-57. https://intjmm.com/index.php/ijmm/article/view/124