The final, formatted version of the article will be published soon.
ORIGINAL RESEARCH article
Front. Plant Sci.
Sec. Technical Advances in Plant Science
Volume 15 - 2024 |
doi: 10.3389/fpls.2024.1496801
This article is part of the Research Topic UAVs for Crop Protection: Remote Sensing, Prescription Mapping and Precision Spraying View all 9 articles
Adaptive Spatial-Channel Feature Fusion and Self-Calibrated Convolution for Early Maize Seedlings Counting in UAV Images
Provisionally accepted- 1 Ningxia University, Yinchuan, China
- 2 Institute of Water Resources Research of Ningxia Hui Autonomous Region, yinchuan, China
Counting crop plants is a laborious and error-prone undertaking in agricultural science, particularly crucial for yield forecasting, field management, and experimental studies. Varying altitude of UAV can affect the dimensions of early maize seedlings in captured images, potentially resulting in blurred features and reduced counting accuracy. To address this issue, we developed a deep learning methodology based on DINO, called RC-Dino, which is specifically designed for the purpose of counting seedlings in UAV-acquired images. The efficacy of our approach was validated using a field-collected dataset of early maize seedlings. Firstly, we introduced a novel self-calibrating convolutional layer, RSCconv, which can be readily incorporated into existing architectural frameworks. The RSCconv is designed to calibrate spatial domain features, thereby improving the representation of early maize seedlings compared to non-seedlings elements within the feature maps. To further enhance the discriminability of early maize seedlings, we have devised a novel module for adaptive spatial feature fusion (ASCFF). This module adaptively fuses feature maps from different layers extracted from the backbone network. Furthermore, we employed transfer learning to integrate the pre-trained weights of the backbone with RSCconv, facilitating accelerated convergence and enhanced accuracy. A dataset, designated the Early Maize Seedlings Dataset (EMSD), was assembled comprising 1,233 images of early maize seedlings, with annotations for a total of 83,404 seedlings. The results of our testing on this dataset demonstrated that our proposed method exhibited the highest performance for the counting of early maize seedlings. In particular, the AP and Recall were increased by 16.29% and 8.19%, respectively, in comparison to the DINO model. In terms of seedling counting, our approach demonstrated a coefficient of determination (R²) in different datasets that was superior to the performance of other models, including Faster R-CNN, RetinaNet, YOLOX, and Deformable DETR. Therefore, our approach is suitable for the accurate counting of early maize seedlings in the field. By integrating RSCconv and ASCFF into the Faster R-CNN, RetinaNet and Deformable DETR models, we have improved their detection and counting accuracy, further validating the effectiveness of our method. The RSCconv and ASCFF code mentioned in this paper can be found here: https://github.com/collapser-AI/RC-Dino.
Keywords: Early maize seedlings counting, Adaptive spatial-channel feature fusion, Self-calibrated convolution, RC-Dino, deep learning
Received: 15 Sep 2024; Accepted: 13 Dec 2024.
Copyright: © 2024 Sun, Yang, Ding, Sun, Sun, Guo and Zhu. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
* Correspondence:
Zhi Yang, Institute of Water Resources Research of Ningxia Hui Autonomous Region, yinchuan, China
Yimin Ding, Ningxia University, Yinchuan, China
Boyan Sun, Ningxia University, Yinchuan, China
Saiju Sun, Ningxia University, Yinchuan, China
Zhen Guo, Ningxia University, Yinchuan, China
Lei Zhu, Ningxia University, Yinchuan, China
Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.