Author
Listed:
- Iman Hindi
- Adham Alsharkawi
- Malik Al-Ajlouni
- Bassam Qarallah
Abstract
Greenhouse climate control is essential for optimizing crop growth while minimizing resource consumption in controlled environment agriculture. Traditional rule-based and fixed-action strategies often struggle to achieve a balance between these objectives. This paper proposes a reinforcement learning (RL) based framework for greenhouse climate control, integrating deep learning models to predict both crop growth and resource consumption. The framework enables an RL agent to optimize greenhouse control setpoints dynamically, maximizing crop yield while ensuring sustainable resource usage. The proposed system incorporates a Multi-Layer Perceptron (MLP) model to predict internal greenhouse climate conditions, a Long Short-Term Memory (LSTM) model for crop parameter estimation, and a separate LSTM model for forecasting daily resource consumption. These models collectively simulate a greenhouse environment where an RL agent learns to regulate temperature, CO2 concentration, and irrigation levels by interacting with the virtual environment. A custom reward function is designed to guide the agent, considering key crop parameters; stem elongation, stem thickness, and cumulative trusses; alongside resource consumption metrics, including heating, electricity, CO2, and irrigation costs. To enhance the adaptability of the RL agent, a feature-selection mechanism identified the most influential climate and control features, reducing observation complexity and accelerating convergence. Retraining under stochastic weather conditions strengthened robustness to dynamic environments, enabling the agent to consistently outperform fixed-action strategies. Evaluation revealed a stable Pareto frontier between yield and resource consumption, confirming that the framework accurately captured the productivity and sustainability trade-off and remained robust across varying reward-weight settings. Comparative analysis of multiple RL algorithms; Proximal Policy Optimization (PPO), Deep Deterministic Policy Gradient (DDPG), Soft Actor-Critic (SAC), and Twin Delayed Deep Deterministic Policy Gradient (TD3) demonstrated that TD3 outperforms other algorithms, achieving the highest cumulative rewards and reaching optimal policies faster. Experimental evaluations demonstrate that the proposed TD3 RL-based greenhouse control system achieves higher crop yield growth rates while optimizing resource usage, outperforming conventional greenhouse control strategies. This study presents a novel data-driven, adaptive greenhouse management approach, bridging the gap between crop growth modeling and autonomous climate control, contributing to sustainable and intelligent agricultural practices.
Suggested Citation
Iman Hindi & Adham Alsharkawi & Malik Al-Ajlouni & Bassam Qarallah, 2026.
"Enhancing autonomous agriculture control systems in greenhouses for sustainable resource usage using deep learning techniques,"
PLOS ONE, Public Library of Science, vol. 21(3), pages 1-47, March.
Handle:
RePEc:plo:pone00:0344946
DOI: 10.1371/journal.pone.0344946
Download full text from publisher
Corrections
All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:plo:pone00:0344946. See general information about how to correct material in RePEc.
If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.
We have no bibliographic references for this item. You can help adding them by using this form .
If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.
For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: plosone (email available below). General contact details of provider: https://journals.plos.org/plosone/ .
Please note that corrections may take a couple of weeks to filter through
the various RePEc services.