IDEAS home Printed from https://ideas.repec.org/a/dba/pappsa/v3y2025ip190-203.html
   My bibliography  Save this article

AI-Driven Computational Resource Optimization: A Hybrid Deep Reinforcement Learning Framework for Enhancing Large-Scale Model Efficiency

Author

Listed:
  • Li, Xiaoying

Abstract

This paper presents an AI-driven approach to computational resource optimization aimed at enhancing efficiency and resource utilization in large-scale AI model training and inference processes. As AI models continue to grow exponentially in size, computational resource demands have increased dramatically, with studies indicating that resource utilization efficiency in large model training and deployment typically falls below 60%, resulting in significant cost waste and energy consumption. The proposed framework leverages reinforcement learning and predictive analytics techniques to implement intelligent resource allocation and task scheduling across di-verse hardware environments (GPUs, CPUs, specialized accelerators). The multi-layered architecture incorporates four core components: resource monitoring, workload prediction, adaptive scheduling, and dynamic optimization, capable of adjusting resource configurations dynamically based on workload characteristics and hardware capabilities. Experimental evaluation across 32 large-scale AI training and inference scenarios demonstrated an average 38.7% throughput improvement and 42.3% energy consumption reduction compared to traditional scheduling meth-odds. Field validation through four industry case studies further confirmed the practical value of this approach in financial services, e-commerce, healthcare, and industrial sectors, achieving an average 31.2% increase in resource utilization and 24.8% reduction in operational costs. Economic analysis indicates substantial return on investment (169% average over 6 months) through improved computational efficiency and reduced infrastructure expenses. These research findings have significant strategic implications for enhancing core competitiveness in high-performance computing and AI infrastructure, contributing to reduced dependency on third-party computational resources, accelerated AI innovation cycles, and advancement of green computing initiatives.

Suggested Citation

Handle: RePEc:dba:pappsa:v:3:y:2025:i::p:190-203
as

Download full text from publisher

File URL: https://pinnaclepubs.com/index.php/PAPPS/article/view/274/283
Download Restriction: no
---><---

More about this item

Keywords

;
;
;
;

Statistics

Access and download statistics

Corrections

All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:dba:pappsa:v:3:y:2025:i::p:190-203. See general information about how to correct material in RePEc.

If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

We have no bibliographic references for this item. You can help adding them by using this form .

If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Joseph Clark (email available below). General contact details of provider: https://pinnaclepubs.com/index.php/PAPPS .

Please note that corrections may take a couple of weeks to filter through the various RePEc services.

IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.