Author
Listed:
- Cardo-Miota, Javier
- Khadem, Shafi
- Bahloul, Mohamed
Abstract
This paper presents a reinforcement learning (RL)-based optimization strategy for minimizing the electricity bill of a residential prosumer equipped with a photovoltaic (PV) system and a battery energy storage system (BESS). Specifically, we implement a Twin Delayed Deep Deterministic Policy Gradient (TD3) agent, an actor–critic RL algorithm with a continuous action space. The proposed model leverages historical PV generation and household consumption data to optimize the charging and discharging schedule of the BESS under a three-period tariff scheme. To enhance the learning process, a Long Short-Term Memory (LSTM) layer is integrated into both the actor and critic networks, allowing the agent to capture temporal dependencies in energy demand and PV generation. The RL agent is trained and evaluated using real-world data from a residential PV installation in Ireland, a region characterized by high variability of solar generation. The results demonstrate that the TD3-based approach effectively reduces electricity bill costs by strategically charging during low-tariff periods and discharging during peak-price hours, achieving an approximate 30% reduction in the monthly bill compared to a PV-only scenario, and a 21% relative to a benchmark strategy. In addition, the TD3 agent is compared to a Deep Deterministic Policy Gradient (DDPG) agent, another RL-based approach for continuous action spaces. The results confirm that TD3 outperforms DDPG in both cost savings and learning stability, validating the effectiveness of our proposed method.
Suggested Citation
Cardo-Miota, Javier & Khadem, Shafi & Bahloul, Mohamed, 2025.
"Deep reinforcement learning based electricity bill minimization strategy for residential prosumer,"
Mathematics and Computers in Simulation (MATCOM), Elsevier, vol. 238(C), pages 296-305.
Handle:
RePEc:eee:matcom:v:238:y:2025:i:c:p:296-305
DOI: 10.1016/j.matcom.2025.06.011
Download full text from publisher
As the access to this document is restricted, you may want to
for a different version of it.
Corrections
All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:eee:matcom:v:238:y:2025:i:c:p:296-305. See general information about how to correct material in RePEc.
If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.
We have no bibliographic references for this item. You can help adding them by using this form .
If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.
For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Catherine Liu (email available below). General contact details of provider: http://www.journals.elsevier.com/mathematics-and-computers-in-simulation/ .
Please note that corrections may take a couple of weeks to filter through
the various RePEc services.