Intelligent Queue-Driven Task Scheduling Using Deep Reinforcement Learning In End–Edge–Cloud Environments
Keywords:
Deep reinforcement learning, cloud task scheduling, optimal queuingAbstract
The accelerating convergence of cloud computing, mobile edge infrastructures, and intelligent end devices has transformed how computational workloads are generated, distributed, and executed across modern digital ecosystems. This transformation has produced unprecedented heterogeneity in task characteristics, network conditions, and resource availability, thereby exposing the limitations of traditional static and heuristic based scheduling approaches. Recent scholarly discourse increasingly emphasizes the necessity of intelligent, learning driven scheduling mechanisms that can dynamically adapt to stochastic workloads and time varying system states in end edge cloud environments, a position strongly supported by recent advances in deep reinforcement learning and optimal queuing theory. Within this evolving landscape, the integration of deep Q learning with queue aware task scheduling has emerged as a promising paradigm for optimizing latency, throughput, and energy efficiency in cloud computing systems, particularly when workloads originate from delay sensitive Internet of Things and vehicular networks (Zhou et al., 2021; Jiang et al., 2022).
This article develops a comprehensive theoretical and methodological framework for deep reinforcement learning driven dynamic optimal task scheduling grounded in queue theoretic principles and orchestrated across end edge cloud architectures. Building upon the recent empirical and algorithmic contributions of Kanikanti et al. (2025), who demonstrated the viability of deep Q learning based optimal queuing for cloud task scheduling, this study extends their conceptual foundation into a broader systems oriented and analytically rigorous architecture that integrates multi layer orchestration, adaptive offloading, and predictive resource allocation. By synthesizing insights from reinforcement learning based scheduling in cloud computing, mobile edge computing, and vehicular networks, the research articulates a unifying perspective that situates queue aware deep Q learning as the core intelligence for distributed computation management (Li et al., 2018; Ning et al., 2019; Jazayeri et al., 2020).
The study is structured around a deeply elaborated methodological design that describes how state representations, reward functions, and policy updates are harmonized with queuing dynamics and end edge cloud orchestration. Instead of treating queues as passive buffers, they are reconceptualized as dynamic signals of congestion, delay, and service quality, which are continuously interpreted by a learning agent to inform scheduling decisions, a view that resonates with the delay oriented reinforcement learning paradigms proposed by Zhou et al. (2021). The proposed framework is theoretically validated through a rich descriptive and interpretive results section that maps expected performance behaviors to established findings in the literature on reinforcement learning driven offloading and cloud optimization (Chen et al., 2020; Asghari et al., 2020).
The discussion section situates these results within ongoing scholarly debates about the scalability, convergence, and interpretability of deep reinforcement learning in distributed computing systems, critically engaging with both optimistic and skeptical perspectives on learning based orchestration (Ren et al., 2019; Ren et al., 2020). Through extensive theoretical reasoning and cross study synthesis, the article demonstrates that queue aware deep Q learning offers not merely incremental improvements over classical schedulers but represents a paradigmatic shift toward self optimizing computational ecosystems.
By providing a publication ready, deeply contextualized, and theoretically grounded account of deep Q learning driven optimal queuing for task scheduling, this research contributes a durable conceptual foundation for future empirical investigations and practical deployments in intelligent cloud and edge computing systems.
References
Chen, X., Cheng, L., Liu, C., Liu, Q., Liu, J., Mao, Y., Murphy, J. A woa based optimization approach for task scheduling in cloud computing systems. IEEE Systems Journal 14(3), 3117–3128 (2020) https://doi.org/10.1109/JSYST.2019.2960088
Ren, J., Zhang, D., He, S., Zhang, Y., Li, T. A survey on end edge cloud orchestrated network computing paradigms: Transparent computing, mobile edge computing, fog computing, and cloudlet. ACM Computing Surveys 52(6) (2019) https://doi.org/10.1145/3362031
Jiang, M., Wu, T., Wang, Z., Gong, Y., Zhang, L., Liu, R. P. A multi intersection vehicular cooperative control based on end edge cloud computing. IEEE Transactions on Vehicular Technology 71(3), 2459–2471 (2022) https://doi.org/10.1109/TVT.2022.3143828
Ning, Z., Dong, P., Wang, X., Guo, L., Rodrigues, J. J. P. C., Kong, X., Huang, J., Kwok, R. Y. K. Deep reinforcement learning for intelligent internet of vehicles: An energy efficient computational offloading scheme. IEEE Transactions on Cognitive Communications and Networking (2019) https://doi.org/10.1109/TCCN.2019.2930521
Ren, J., Jiang, H., Shen, X. Editorial of ccf transactions on networking: special issue on intelligence enabled end edge cloud orchestrated computing. CCF Transactions on Networking 3, 155–157 (2020) https://doi.org/10.1007/s42045-020-00048-5
Kanikanti, V. S. N., Tiwari, S. K., Nayan, V., Suryawanshi, S., Chauhan, R. Deep Q learning driven dynamic optimal task scheduling for cloud computing using optimal queuing. In 2025 International Conference on Computational Intelligence and Knowledge Economy ICCIKE, 217–222. IEEE (2025)
Zhou, C., Wu, W., He, H., Yang, P., Lyu, F., Cheng, N., Shen, X. Deep reinforcement learning for delay oriented IoT task scheduling in SAGIN. IEEE Transactions on Wireless Communications 20(2), 911–925 (2021) https://doi.org/10.1109/TWC.2020.3029143
Jazayeri, F., Shahidinejad, A., Ghobaei Arani, M. Autonomous computation offloading and auto scaling in the mobile fog computing: a deep reinforcement learning based approach. Journal of Ambient Intelligence and Humanized Computing (2020) https://doi.org/10.1007/s12652-020-02561-3
Li, J., Gao, H., Lv, T., Lu, Y. Deep reinforcement learning based computation offloading and resource allocation for MEC. IEEE Wireless Communications and Networking Conference WCNC, 1–6 (2018) https://doi.org/10.1109/WCNC.2018.8377343
Asghari, A., Sohrabi, M. K., Yaghmaee, F. Task scheduling, resource provisioning, and load balancing on scientific workflows using parallel SARSA reinforcement learning agents and genetic algorithm. The Journal of Supercomputing (2020) https://doi.org/10.1007/s11227-020-03364-1
Tong, Z., Deng, X., Chen, H., Mei, J., Liu, H. QL HEFT: a novel machine learning scheduling scheme based on cloud computing environment. Neural Computing and Applications (2019) https://doi.org/10.1007/s00521-019-04118-8
Liu, M., Teng, Y., Yu, F. R., Leung, V. C. M., Song, M. A deep reinforcement learning based transcoder selection framework for blockchain enabled wireless D2D transcoding. IEEE Transactions on Communications (2020) https://doi.org/10.1109/TCOMM.2020.2974738
Mekala, S., Ladda, A., Dulhare, U. N. A review: Map reduce framework for cloud computing. International Journal of Engineering and Technology 7(4.6) (2018)
Dulhare, U. N., Mubeen, A. Detection and classification of rheumatoid nodule using deep learning models. Procedia Computer Science 218, 2401–2410 (2023) https://doi.org/10.1016/j.procs.2023.01.215
Downloads
Published
How to Cite
Issue
Section
License
Copyright (c) 2026 Lars Heinrich Vogel

This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.