This is an unedited manuscript accepted for publication and provided as an Article in Press for early access at the author’s request. The article will undergo copyediting, typesetting, and galley proof review before final publication. Please be aware that errors may be identified during production that could affect the content. All legal disclaimers of the journal apply.
Pratik Danodia,
Kamlesh Lakhwani,
Bhavna Sharma,
- Student, Department of CSE, JECRC University, Jaipur, Rajasthan, India
- Professor, Department of CSE, JECRC University, Jaipur, Rajasthan, India
- Associate Professor, Department of CSE, JECRC University, Jaipur, Rajasthan, India
Abstract
Modern distributed and heterogeneous computing systems face significant challenges in dealing with dynamically changing workloads, resource fragmentation, and changing latencies; existing traditional, or rule-based, schedulers are no longer useful in achieving the best system performance. Such limitations highlight the importance of the adaptive scheduling paradigms that are able to learn, to forecast and reaction to the real red conditions in the system. The middleware of artificial-intelligence is also an attractive solution to this issue, as it allows to incorporate machine-learning and reinforcement-learning algorithms in the orchestration layer and, therefore, promote intelligent decision-making and improve resource distribution and tasks assignment in a progressive fashion. This paper suggests a flexible scheduling system, which runs on the AI middleware, incorporates workload forecasting models, reinforcement-learning-based scheduling policies, reactive scaling, and energy-sensitive decision making. In the approach, there would be a multi-layer middleware design, which would incorporate real time monitoring, state- action-reward learning, and feedback-based scheduler. The major contributions of this paper include the creation of an RL-supported adaptive scheduling algorithm, a single middleware architecture that can be applied in distributed settings and an in-depth performance analysis covering cloud and edge computing contexts. The experimental results show that the schedulers achieve considerable benefits in delay, throughput, resource exploitation and energy efficiency when compared with the baseline schedulers. This can apply to the cloud platform environments, IoT-edge environments and enterprise distributed systems hence providing a scalable and smart solution to modern workload-management issues.
Keywords: Adaptive Scheduling, AI Middleware, Resource Optimization, Distributed Systems, Edge Computing, Workload Management
Pratik Danodia, Kamlesh Lakhwani, Bhavna Sharma. Adaptive Task Scheduling And Resource Optimization Using Ai Middleware. Recent Trends in Parallel Computing. 2026; 13(01):-.
Pratik Danodia, Kamlesh Lakhwani, Bhavna Sharma. Adaptive Task Scheduling And Resource Optimization Using Ai Middleware. Recent Trends in Parallel Computing. 2026; 13(01):-. Available from: https://journals.stmjournals.com/rtpc/article=2026/view=242302
References
- Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser Ł, Polosukhin I. Attention is all you need. Advances in neural information processing systems. 2017;30.
- Brown T, Mann B, Ryder N, Subbiah M, Kaplan JD, Dhariwal P, Neelakantan A, Shyam P, Sastry G, Askell A, Agarwal S. Language models are few-shot learners. Advances in neural information processing systems. 2020;33:1877-901.
- Houlsby N, Giurgiu A, Jastrzebski S, Morrone B, De Laroussilhe Q, Gesmundo A, Attariyan M, Gelly S. Parameter-efficient transfer learning for NLP. InInternational conference on machine learning 2019 May 24 (pp. 2790-2799). PMLR.
- Lester B, Al-Rfou R, Constant N. The power of scale for parameter-efficient prompt tuning. InProceedings of the 2021 conference on empirical methods in natural language processing 2021 Nov (pp. 3045-3059).
- Hu EJ, Shen Y, Wallis P, Allen-Zhu Z, Li Y, Wang S, Wang L, Chen W. Lora: Low-rank adaptation of large language models. Iclr. 2022 Apr 25;1(2):3.
- Dettmers T, Pagnoni A, Holtzman A, Zettlemoyer L. Qlora: Efficient finetuning of quantized llms. Advances in neural information processing systems. 2023 Dec 15;36:10088-115.
- Hasan MM, Islam MM. High-Performance Computing Architectures For Training Large-Scale Transformer Models In Cyber-Resilient Applications. ASRC Procedia: Global Perspectives in Science and Scholarship. 2022 Apr 29;2(1):193-226.
- Zhao L, Gao W, Fang J. Optimizing large language models on multi-core CPUs: a case study of the BERT model. Applied Sciences. 2024 Mar 11;14(6):2364.
- Han Z, Gao C, Liu J, Zhang J, Zhang SQ. Parameter-efficient fine-tuning for large models: A comprehensive survey. arXiv preprint arXiv:2403.14608. 2024 Mar 21.
- Mao H, Schwarzkopf M, Venkatakrishnan SB, Meng Z, Alizadeh M. Learning scheduling algorithms for data processing clusters. InProceedings of the ACM special interest group on data communication 2019 Aug 19 (pp. 270-288).

Recent Trends in Parallel Computing
| Volume | 13 |
| 01 | |
| Received | 11/02/2026 |
| Accepted | 23/03/2026 |
| Published | 30/04/2026 |
| Publication Time | 78 Days |
Login
PlumX Metrics