Author
Listed:
- Xinying Zhou
(School of Computer Science and Technology, Xinjiang University, Urumqi 830017, China
School of Agricultural Engineering, Jiangsu University, Zhenjiang 212013, China)
- Wenming Chen
(School of Agricultural Engineering, Jiangsu University, Zhenjiang 212013, China)
- Xinhua Wei
(School of Agricultural Engineering, Jiangsu University, Zhenjiang 212013, China
Key Laboratory for Theory and Technology of Intelligent Agricultural Machinery and Equipment of Jiangsu University, Zhenjiang 212013, China)
Abstract
To satisfy the obstacle avoidance requirements of unmanned agricultural machinery during autonomous operation and address the challenge of rapid obstacle detection in complex field environments, an improved field obstacle detection model based on YOLOv8 was proposed. This model enabled the fast detection and recognition of obstacles such as people, tractors, and electric power pylons in the field. This detection model was built upon the YOLOv8 architecture with three main improvements. First, to adapt to different tasks and complex environments in the field, improve the sensitivity of the detector to various target sizes and positions, and enhance detection accuracy, the CBAM (Convolutional Block Attention Module) was integrated into the backbone layer of the benchmark model. Secondly, a BiFPN (Bi-directional Feature Pyramid Network) architecture took the place of the original PANet to enhance the fusion of features across multiple scales, thereby increasing the model’s capacity to distinguish between the background and obstacles. Third, WIoU v3 (Wise Intersection over Union v3) optimized the target boundary loss function, assigning greater focus to medium-quality anchor boxes and enhancing the detector’s overall performance. A dataset comprising 5963 images of people, electric power pylons, telegraph poles, tractors, and harvesters in a farmland environment was constructed. The training set comprised 4771 images, while the validation and test sets each consisted of 596 images. The results from the experiments indicated that the enhanced model attained precision, recall, and average precision scores of 85.5%, 75.1%, and 82.5%, respectively, on the custom dataset. This reflected increases of 1.3, 1.2, and 1.9 percentage points when compared to the baseline YOLOv8 model. Furthermore, the model reached 52 detection frames per second, thereby significantly enhancing the detection performance for common obstacles in the field. The model enhanced by the previously mentioned techniques guarantees a high level of detection accuracy while meeting the criteria for real-time obstacle identification in unmanned agricultural equipment during fieldwork.
Suggested Citation
Xinying Zhou & Wenming Chen & Xinhua Wei, 2024.
"Improved Field Obstacle Detection Algorithm Based on YOLOv8,"
Agriculture, MDPI, vol. 14(12), pages 1-26, December.
Handle:
RePEc:gam:jagris:v:14:y:2024:i:12:p:2263-:d:1540936
Download full text from publisher
Corrections
All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jagris:v:14:y:2024:i:12:p:2263-:d:1540936. See general information about how to correct material in RePEc.
If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.
We have no bibliographic references for this item. You can help adding them by using this form .
If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.
For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .
Please note that corrections may take a couple of weeks to filter through
the various RePEc services.