
94% of researchers rate our articles as excellent or good
Learn more about the work of our research integrity team to safeguard the quality of each article we publish.
Find out more
ORIGINAL RESEARCH article
Front. Plant Sci. , 30 January 2025
Sec. Technical Advances in Plant Science
Volume 15 - 2024 | https://doi.org/10.3389/fpls.2024.1499896
Extracting the navigation line of crop seedlings is significant for achieving autonomous visual navigation of smart agricultural machinery. Nevertheless, in field management of crop seedlings, numerous available studies involving navigation line extraction mainly focused on specific growth stages of specific crop seedlings so far, lacking a generalizable algorithm for addressing challenges under complex cross-growth-stage seedling conditions. In response to such challenges, we proposed a generalizable navigation line extraction algorithm using classical image processing technologies. First, image preprocessing is performed to enhance the image quality and extract distinct crop regions. Redundant pixels can be eliminated by opening operation and eight-connected component filtering. Then, optimal region detection is applied to identify the fitting area. The optimal pixels of plantation rows are selected by cluster-centerline distance comparison and sigmoid thresholding. Ultimately, the navigation line is extracted by linear fitting, representing the autonomous vehicle’s optimal path. An assessment was conducted on a sugarcane dataset. Meanwhile, the generalization capacity of the proposed algorithm has been further verified on corn and rice datasets. Experimental results showed that for seedlings at different growth stages and diverse field environments, the mean error angle (MEA) ranges from 0.844° to 2.96°, the root mean square error (RMSE) ranges from 1.249° to 4.65°, and the mean relative error (MRE) ranges from 1.008% to 3.47%. The proposed algorithm exhibits high accuracy, robustness, and generalization. This study breaks through the shortcomings of traditional visual navigation line extraction, offering a theoretical foundation for classical image-processing-based visual navigation.
Sugarcane (Saccharum officinarum) is a vital sugar-producing and energy-producing crop widely cultivated in tropical and subtropical countries or regions (Wang W. et al., 2023), with an annual global yield of approximately 1.9 billion tonnes (FAO, 2022). Diverse field management tasks at the seedling stage, such as stumping, weeding, spraying pesticides, and root-zone soil backfill, are crucial in improving the sugarcane yield and quality. At present, these tasks severely rely on manual operation. In the face of the aging labor population and the increase in labor costs, it is urgently needed to explore intelligent agricultural machinery for field management. Intelligent agricultural machinery can be a potential resolution to perform field management tasks, and navigation is one of the key technologies to achieve intelligent detection/control of agricultural machinery.
Commonly, agricultural machinery navigation methods mainly include Global Navigation Satellite System (GNSS) navigation (Luo et al., 2009; Liu et al., 2018; Zhang et al., 2022), radar navigation (Blok et al., 2019), ultrasonic navigation (Chen et al., 2018) and visual navigation (Liu et al., 2024; Bai et al., 2023; Shi et al., 2023). According to the available literature, visual navigation attracted more interest in these methods. This phenomenon might occur because visual navigation, relying on the interpretation and judgment of various landscapes, can be effective in diverse settings. This includes extensive, continuous farmlands on plains and fragmented, small-scale farmlands in hilly and mountainous terrains.
It is well known that the visual navigation of crop seedlings can be classified into two categories: (I) the navigation based on the ridge centerline and (II) the navigation based on crop plantation-row centerlines. For the first category, in the early years, one typically representative method is the least-square-based navigation line detection algorithm of tomato ridge (Wang et al., 2012). In recent years, further advances have been found in this regard. For instance, Yang et al. (2020) proposed a method of real-time extraction of the navigation line between the corn rows based on a dynamic Region of Interest. Facing the higher corn seedlings, Gong et al. (2020, 2022) gave two corn field navigation line extraction methods by combining edge detection and area localization and integrating gradient descent and corner detection. Chen et al. (2020) investigated a median point Hough transformation algorithm to fit the navigation paths of greenhouse tomato-cucumber seedlings. Following this previous study, Chen et al. (2021) also investigated a prediction-point Hough transform to extract the navigation path for greenhouse cucumber seedlings. With the in-depth application of deep learning algorithms, ever-increasing researchers have adopted deep learning algorithms based on semantic segmentation technology to obtain navigation lines between the crop rows. For example, Rao et al. (2021) combined the U-net and Fast-Unet models to generate the navigation line. Zhao et al. (2021) used remote sensing images acquired from unmanned aerial vehicles and extracted corn field ridge centerline based on a fully convolutional network. Li et al. (2022a) achieved the navigation line extraction using a Faster-U-net model and adaptive multi-ROIs. Song et al. (2022) proposed a navigation algorithm based on semantic segmentation in wheat fields using an RGB-D camera. Yang et al. (2023) develops a real-time crop row detection algorithm for corn fields, leveraging YOLO neural network and ROI extraction to achieve high accuracy and robustness. For such a category, it is evident that navigation utilizing the ridge centerline depends on feature point extraction between the adjacent crop rows and the color difference in ridge and plantation rows. The limited condition of non-intersecting overlapping leaves between adjacent crop rows compounds this reliance on either classical image processing or deep learning algorithms. Especially for the semantic segmentation-based ridge centerline extraction, from the point of view of real-world application, it remains challenging to integrate with downstream low-cost, lightweight edge devices after obtaining semantic segmentation results.
As for the second category, navigation based on crop plantation-row centerlines has been a research hotspot in the past two decades. Most of the available literature in plantation-row centerlines extraction were based on the development of classical image processing algorithms, such as Hough transform based on connected component labeling (Rao and Ji, 2007), improved Hough transformation (Zhao et al., 2009; Diao et al., 2015; Wang et al., 2020), median-point Hough transform (Li et al., 2022b), least square method (Si et al., 2010; Ma et al., 2022), color model and nearest neighbor clustering algorithm (Zhang et al., 2012), scan-filter algorithm (Li et al., 2013), constraint of liner correlation coefficient (Meng et al., 2013), improved genetic algorithm (Meng et al., 2014), boundary detection and scan-filter (He et al., 2014), multi-ROIs (Jiang et al., 2015; Yang et al., 2020; Yu et al., 2021; Wang A. et al., 2021; Li et al., 2021), SUSAN corner (Zhang et al., 2015), Harris corner points (Zhai et al., 2016a), Census transformation (Zhai et al., 2016b), particle swarm optimization (Meng et al., 2016), SUSAN corner and improved sequential clustering algorithm (Zhang et al., 2017), image characteristic point and particle swarm optimization-clustering algorithm (Jiang et al., 2017), accumulation threshold of Hough transformation (Chen et al., 2019), sub-regional feature points clustering (Liao et al., 2019), Gaussian heatmap (He et al., 2022), contour computing with vertical projection (He et al., 2021), regional growth and mean shift clustering (Wang A. et al., 2021), binocular vision and adaptive Kalman filter (Zhai et al., 2022), RANSAC (Random Sample Consensus) algorithm (Li et al., 2022c), and region growth sequential clustering-RANSAC algorithm (Fu et al., 2023). These numerous literature revealed that both past and present, a significant proportion of researchers are still keen to perform navigation centerline extraction based on classical image processing algorithms. Classical image processing algorithms can be efficiently run on CPUs without the requirement for high-performance computing resources such as GPUs. When handling data from diverse sources and with varying characteristics, the compatibility and consistency of heterogeneous systems pose significant challenges. Classical image processing algorithms, due to their simplicity, are adaptable to various affordable systems and application scenarios, simplifying the complexities associated with managing heterogeneous data. Moreover, these algorithms are more explainable and easier to implement in low-performance computing platforms. Nevertheless, when facing complex image processing, some algorithm executions are time-consuming and can result in a limited generalization ability.
With the further popularity of deep learning technology, some researchers have attempted to develop various deep learning algorithms to extract navigation centerlines in the last few years to tackle this challenge. For instance, Lin et al. (2020) used the Faster R-CNN to detect the transplanting rice seeding in complex paddy field environments, obtaining mean absolute errors of a deviation of 8.5 mm in the lateral distance and 0.50° of travel angle. Zhang et al. (2020) extracted the centerlines of rice seedlings based on the YOLOv3 model, achieving a mean average precision of 91.47%, a mean average angle error of 0. 97° in the extracted centerline, and an average runtime of 82. 6 ms for one image. Diao et al. (2024) developed an algorithm for recognizing corn crop rows at different growth stages using the ST-YOLOv8s network. Their approach, which included a swin transformer-based backbone and a local–global detection method, demonstrated improved mean average precision (MAP) and accuracy compared to YOLOv5s, YOLOv7 and YOLOv8s, while reducing the average angle error and fitting time in crop row detection experiments. Li et al. (2023) presents E2CropDet, an efficient end-to-end deep learning model for crop row detection. Zheng and Wang (2023) used Pix2Pix Net to improve navigation line extraction ability in image pre-processing. Ruan et al. (2023) proposed a crop row detection method based on YOLO-R, attaining accuracy values of 93.91%, 95.87%, and 89.87% on the seven-day, 14-day, and 21-day for rice seedling row detection, respectively. Diao et al. (2023) developed a navigation line extraction algorithm based on an improved YOLOv8s network in corn seedlings, getting a significantly enhanced performance in mean average precision and F1, as compared to the YOLOv7 network and original YOLOv8 network. Liu et al. (2023) showed a recognition method of corn crop rows using the MS-ERFNet model, the mean intersection over union (mIoU) and the pixel accuracy (PA) of the MSERFNet model was 93.40% and 97.54%, respectively, which were higher than other models. Wang S. et al. (2021, 2023a, 2023b) adopted row vector grid classification, initial clustering, and exterior point elimination, as well as sub-region growth and outlier removal, to recognize straight or curved rice seedling rows, comprehensively realizing the centerline extraction of transplanted rice seedling rows under different paddy field environments.
Undoubtedly, deep learning algorithms in crop seedling centerline extraction require less execution time and showcase a better generalization ability. Despite these advantages, deep-learning-based methods require comprehensive datasets, considerable image labeling, innovative algorithm development, and rigorous parameter-adjusting processes. In addition, most of the deep learning algorithms lacked verifications in other crop seedlings; as a result, their generalization ability remains limited.
In conclusion, existing approaches often target specific crops, such as rice and corn, which typically exhibit uniformity due to agronomic influences, yet a generalizable navigation line extraction algorithm capable of adapting to various crops, growth stages, and environmental conditions remains absent. As a result, algorithms tailored to these crops may struggle with generalizability when applied to others, such as sugarcane, which has more complex and diverse characteristics. For instance, in sugarcane seedlings, ridge structure and spacing vary significantly, and little literature has explored this field. A noteworthy exception is Yang et al. (2022), who used LiDAR to extract navigation lines between sugarcane ridges. Moreover, traditional methods and many deep-learning-based approaches focus on site-specific, crop-specific, or growth-stage-specific ridge centerline navigation or plantation-row centerline navigation. A generalizable navigation line extraction algorithm that can handle cross-site, cross-crop, and cross-growth-stage scenarios has yet to be reported. Addressing this gap, our study emphasizes the development of an algorithm designed with broader applicability, aiming to bridge the limitations of the current literature.
In this study, a generalizable navigation line extraction algorithm based on classical image processing technologies was proposed. This study aims to extract the navigation line for cross-growth-stage sugarcane plantation rows under complex in-field environments using classical image processing technologies. The specific objectives of the study were to (a) propose an adaptive navigation line extraction algorithm of sugarcane plantation rows in different early growth stages, (b) validate the effectiveness of the proposed algorithm in complex sugarcane field environments, and (c) apply the proposed algorithm into the navigation line extraction of plantation rows in rice and corn, verifying its generalization capability.
The proposed navigation line extraction method comprises four steps: image acquisition, image preprocessing, optimal region detection, and navigation line fitting.
The image acquisition location of sugarcane seedlings is situated in Guangxi subtropical new town of agri-forestry sciences, which belongs to Guangxi University, Guangxi Province, China (22.52°N, 107.79°E). The sugarcane seedling images were taken using a mobile device, Huawei Mate 50, positioned at a declination angle of about 45° relative to the terrestrial surface on May 2023. In total, 153 images were captured at different times.
Corn seedlings’ image acquisition location is in Guangxi University Experimental Base, Guangxi Province, China (22.86°N, 108.30°E). The corn seedling images were taken using the same Huawei Mate 50, positioned at a declination angle of about 45° relative to the terrestrial surface on August 2023. In total, 218 images were captured at different times.
The rice seedlings’ image acquisition location is in Guangxi University Experimental Base, Guangxi Province, China (22.86°N, 108.30°E). The rice seedling images were taken using the same Huawei Mate 50, positioned at a declination angle of 45° relative to the terrestrial surface on August 2023. In total, 135 images were captured at different times.
These acquired images cover diverse in-field conditions such as varying light, inter-row spacing variability, decay branches, overlapping leaves, soil color difference, and disordered vegetation, as well as multiple growth stages which are determined by the number of crop leaves such as ‘Code 12’ to ‘Code 19’ (Zadoks et al., 1974). To further expand the dataset and evaluate the generalization performance of the algorithm, this study employed the open-source image augmentation library Albumentations to enhance the dataset images. These augmentation operations simulated the distributions of sugarcane, corn, and rice seedlings under various low-light conditions and rainy scenarios.
In the following subsections, the sugarcane seedling images were first selected for algorithm development, while the other two image datasets were used for evaluating their generalization performance, as shown in sections 3.2 to 3.3.
The Windows 11 operating system was used to perform image processing tasks on a notebook with an AMD R7-7745HX and NVIDIA GeForce RTX 4060 with 8 GB VRAM. The algorithm in this paper was implemented in C++ using Microsoft Visual Studio 2022 as the integrated development environment. The flowchart of image processing is shown in Figure 1. These steps will be described further in the following sections.
Figure 1. The overall architecture of image processing. Colored boxes indicate the detailed algorithms used for each key step.
To locate crop rows, the excess green index (ExG) is used to segment the vegetation and non-vegetation areas in the image based on the distinct difference in the green index (Woebbecke et al., 1995). The ExG is defined in Equation 1. Grayscale images can be obtained by applying the ExG, as shown in Supplementary Figure S1B.
where denotes the gray level of the pixel at the position and are determined by the normalized values of the corresponding RGB channels respectively, as shown in Equation 2.
Threshold segmentation using the Otsu method (Otsu, 1979) on grayscale images effectively extracts regions of interest. Let be the proportion of pixels for each grayscale level compared to the total number of pixels in the image, as shown in Equation 3. Let be the threshold segmenting the image into foreground and background. Let and represent the probabilities of foreground and background, respectively. Let and represent the average grayscale values of the foreground and background, respectively. Let be the average grayscale values of the image. Thus, the variance between the two categories can be calculated by the Equations 3 and 4:
where and represent the sum of pixels for each grayscale level and total pixels of the image, respectively, while , , , and . The threshold that maximizes the inter-class variance can be calculated by the Equation 5:
where represents the optimal threshold.
Once the threshold that corresponds to the maximum inter-class variance is defined, it becomes the optimal threshold for the segmentation. The binary image illustrated in Figure 2A can be obtained based on the following Equation 6:
Figure 2. Threshold segmentation and morphological operation. (A) Binary images at different growth stages: (Aa) Slight occlusion, NZPR=11.84%; (Ab) Moderate occlusion, NZPR=22.32%; (Ac) Severe occlusion, NZPR=36.38%; (B) Images after morphological operation: the opening operation was applied in (Ba), (Bb) and (Bc) based on NZPR.
where , in which is a proportion coefficient. According to pre-test, is an optimal proportion coefficient to extract the green features of sugarcane to the binary image.
Given that non-zero pixels in the binary image represent crops, this study proposed a concept of non-zero pixel ratio (NZPR), which denotes the proportion of non-zero pixels relative to the total number of pixels in the image to describe the different growth stages of crops, as Equation 7:
Furthermore, it is observed that the sugarcane seedlings exhibit strong branching ability of their leaves from the seedling stage (Figure 2A). As the growth stage advances, the branching density and coverage of the leaves increase continuously, resulting in severe occlusion of the leaf canopy of adjacent crop rows (Figure 2Ac), which leads to unfavorable repercussions for subsequent clustering processes.
It is worth noting that different agricultural scenarios have varying requirements regarding image resolution. Low-resolution images, which do not require excessive image details, offer better real-time performance. On the other hand, high-resolution images provide more precise information for tasks such as planting and weed control. Given common resolutions used in typical cameras, the image acquired with a resolution of 4096×3072 is converted into two-type images with 640×480 and 1920×1080. As a result, established mathematical models were developed to correlate the two different resolutions—640×480 and 1920×1080—with the application of morphological operations.
To effectively tackle the severe occlusion of the leaf canopy of adjacent crop rows, a morphological operation (erosion followed by dilation) was implemented to eliminate extraneous branches and slight noise. According to pre-test, the iterations of erosion and dilation are determined by our defined NZPR ranges, as shown as below.
1. For images with a resolution of 640×480, its can be described as
2. For images with a resolution of 1920×1080, its can be described as
Compared to linear relationships, nonlinear relationships are more effective in balancing the preservation of image details with the suppression of unnecessary pixels. However, in cases where the NZPR is less than 6%, there are not enough feature pixels in the image, and therefore, erosion is not necessary.
As sugarcane seedlings grow, the NZPR value steadily increases (Figure 2), requiring more corrosion and expansion iterations, leading to more significant differences between them. As shown in Figure 2B, the main stems of sugarcane seedlings were retained in the binary image after the morphological operation. Specifically, a 3×3 kernel was used in this process.
The morphological operations effectively suppressed extraneous branches and leaves in the binary image. However, it was also observed that numerous disconnected noises were introduced into the image, resulting from erosion and dilation. At the same time, the relatively larger connected components constitute the significant parts of the crop.
Based on the observed difference mentioned above, an eight-connected labeling algorithm that considers the pixel neighborhood in all eight directions was applied to identify individual objects. The specific steps for the labeling process are described as follows:
Step 1. Initialize a counter that indicates the count of currently discovered connected components. Initialize a label image to store the label of each pixel.
Step 2. Iterate through each pixel in the input image. Based on the grayscale value of the current pixel, classify it into foreground and background. The presence of a non-zero grayscale value denotes the membership of a pixel to the foreground and, in some cases, to a particular connected component. Identify the labelled pixels within its neighborhood and place them into a set.
Step 3. Identify the smallest element in the set, corresponding to the minimum label value, and assign it to a variable. This variable represents the label of the connected component to which the current pixel might belong. If the set is empty, it indicates that the current pixel has no labelled neighboring pixels, signifying it as the starting point of a newly discovered connected component. Therefore, increment the counter by one and assign this value as the label of the current pixel. If the set is not empty, assign the variable mentioned above as the label of the current pixel.
Step 4. Replace different labels belonging to the same connected component with a single label, referred to as merging equivalence class (MEC). MEC ensures that each connected component has a unique label.
The corresponding algorithm for the connected component labeling process is presented in Algorithm 1.
Algorithm 1. The algorithm of the labeling process.
Furthermore, a pixel re-binarization operation, such as in Equation 11, was employed to eliminate small connected components considered residual pixels after performing morphological opening. Figure 3A illustrates the image that was processed.
Figure 3. Connected component detection and vertical projection. (A) Eight-connected component filtering image; (B) Vertical projection curve from the filtering image, the value of (k) depends on the maximum column pixel value of each image.
where represents the area of the connected component to which the current pixel belongs, represents the mean area of all connected components, denotes a given coefficient, and pre-test has shown that a value of 0.7 achieves better elimination effects.
To extract the center row of sugarcane seedlings in the image, a baseline is needed to predict its potential position. The baseline detection is based on the following two assumptions: 1) In the image (Figure 3A), the crop rows present a trend of being smaller at the top and more prominent at the bottom and leaning towards the center position, while the crop rows at the center position have a smaller or almost vertical inclination. 2) Vertically projecting the image exhibits a specific Gaussian distribution pattern. The maximum column pixel value position is typically located near the central crop row. The central 35% of the image is set as the region of interest (ROI) to obtain accurate results. The column pixel values within this region are then analyzed. The vertical line corresponding to the maximum column pixel value is considered the baseline, as shown in Figure 3B.
This study employed a density-based clustering method to group pixels that belong to the same crop or specific crop row, as shown in Figures 4A, D.
Figure 4. Images of clusters based on density: (A) Illustration of DBSCAN; (B) 2D space partitioning based on KDTree, each vertical and horizontal line represents a split along the x or y axis, respectively, the labels (L1, L2, etc.) indicate the hierarchical levels of the partition; (C) KDTree structure; (D) Indexes 1 to 17 indicate clusters of that density level.
Before the implementation of clustering, it was observed that the number of pixels presented in high-resolution binary images was quite substantial, resulting in a significant utilization of computational resources and time. From the pixel reduction perspective, a sliding window method was applied to reduce unnecessary pixels, save computational time, and increase data processing speed. This method involves computing the average coordinates of all pixels within the window, as shown in Equations 12 and 13, subsequently replacing the original pixels with the resulting mean values. The sliding window moves by its step size at each iteration, covering the entire image. Through the implementation of this method, image processing is becoming more streamlined and efficient.
where and represent the width and height of the sliding window respectively, they are both set to eight in this step.
Ester et al. (1996) proposed the DBSCAN (Density-Based Spatial Clustering of Applications with Noise) algorithm. We considered two critical parameters in the algorithm, the neighborhood radius (epsilon) and the minimum number of neighbors (minPts). Based on the pre-test, the parameters were configured to 200 and 4 respectively.
The traditional DBSCAN algorithm faces computational bottlenecks, particularly when processing high-resolution images. The computational complexity of finding neighboring points increases significantly with the amount of pixel information, leading to longer processing times. To address this, the KD-Tree was introduced as a spatial indexing structure to accelerate neighbor searches in the DBSCAN algorithm, reducing the computational complexity from O(n2) to O(n log n). The KD-Tree implementation, as shown in Figures 4B, C, was further enhanced with OpenMP parallelization to reduce processing time for high resolution images. By leveraging OpenMP, the neighbor search process was divided into smaller tasks distributed across multiple threads, significantly improving computational efficiency. The specific steps for the enhanced DBSCAN process are described as follows:
Step 1: Build a KD-tree using the dataset to accelerate neighbor searches. For each data point, find all points within its epsilon-neighborhood using the KD-tree and cache these neighbors for quick access. Perform parallel neighbor searches using OpenMP.
Step 2: Initialize all data points as unclassified by setting their cluster IDs to -1.
Step 3: For each unclassified data point, perform the following:
A. Retrieve its cached epsilon-neighborhood from the KD-tree.
B. If the number of neighbors (including the point itself) is less than the user-defined minimum (minPts), then:
● Mark the data point as noise (remain unclassified).
C. Else if the number of neighbors is greater than or equal to minPts, then:
● Assign the data point to a new cluster by setting its cluster ID to a new unique value.
● Initialize a queue and add the current data point to it to begin expanding the cluster.
Step 4: While the queue is not empty, perform the following to expand the cluster:
A. Dequeue a data point from the queue.
B. Retrieve its cached epsilon-neighborhood.
C. If the number of neighbors is greater than 1 (meaning the point has at least one neighbor besides itself), then:
● For each neighbor in the epsilon-neighborhood:
1. If the neighbor is unclassified (cluster ID is -1), then:
a. Assign it to the current cluster by setting its cluster ID to the current cluster ID.
b. Enqueue the neighbor to process its neighbors in subsequent iterations.
Step 5: Repeat Steps 3 and 4 for all data points until all points are classified either into clusters or marked as noise.
The corresponding algorithm for the process mentioned above is presented in Algorithm 2.
Algorithm 2. The algorithm of DBSCAN using KD-tree acceleration.
The plantation row that is required to fit the navigation line is considered the centerline in Figure 3A. By conducting a comparative analysis of the distances of clusters to the baseline depicted in Figure 3B and subsequently delineating a specific ROI, a more precise estimation of the centerline can be obtained, as shown in Figure 5C.
Figure 5. Selection of median row: (A) Compare with ; (B) Detect the presence of clusters to verify if there is at least one point within the region of interest; (C) Obtain the centerline.
The comparative analysis compares the distance from each cluster centroid to the baseline with the mean distance of all cluster centroids to the baseline, as shown in Equation 14. This approach enables the extraction of potential target clusters while filtering out clusters that exhibit a minimal proportion of areas within the ROI, yet the centroid is distant from this region, as illustrated in Figure 5A.
where represents an absolute distance from - cluster centroid of its x-coordinate to the baseline, is calculated as the mean of all centroid distances and the coefficient of distance (cod) will start with an initial value of 0.65 and increase by 0.05 in each iteration until at least a cluster that satisfies Equation 14 is found.
However, not all clusters that satisfy Equation 14 are necessarily relevant to the required fitting clusters. These clusters may encompass redundant significant connected components and certain biased positional crops. To address this issue, an isosceles trapezoidal ROI with the baseline as its median was employed. Figure 5B illustrates the ROI, which consists of two oblique lines, each spanning from the top to the bottom, with a length equivalent to 0.07 times the image width. Furthermore, the slopes of these two lines have an absolute value of 1/12. The cross-product operation, as shown in Equation 15, was applied to assess the position of the cluster relative to the oblique lines. Equation 16 determines whether the cluster is within the ROI. The clusters contributing to a non-zero value in Equation 16 are regarded as the desired median row, as illustrated in Figure 5C.
where and represent two points on the oblique line, represents the point which the relative position is to be determined. The number of points within the two oblique lines is given by:
where and determine the left oblique line, and determine the right oblique line, , and is the indicator function, when is true, ; when is false, .
In crop analysis, accurately selecting a segmentation representing the crop row is essential for fitting a navigation line. Despite the series mentioned above of processing steps, it was observed that the centerline depicted in Figure 6A may still contain discrete branch portions, which can lead to a decrease in fitting accuracy. To mitigate this issue, a novel thresholding technique is proposed in this study, which utilizes a sigmoid function based on NZPR to segment centerline images.
Figure 6. Sigmoid thresholding process. (A) Median rows at different growth stages: (Aa) NZPR=1.84%, =0.95; (Ab) NZPR=8.96%, =0.91; (Ac) NZPR=22.32%, =0.76; (B) Vertical projection curves of centerlines: the horizontal line is applied in (Ba), (Bb) and (Bc) to determine the blue area which corresponds to the final fitting area; (C) Centerlines after segmentation, the larger the NZPR, the more likely it is to segment a smaller fitting area.
The threshold for segmenting is determined by Equation 17, which maps different NZPR values to ranges from 0 to 1, normalizing pixel values in the image. Images with higher NZPR values imply later growth stages with denser median crop rows, particularly in the central stem region. Consequently, there is an urgent need for segmentation in such areas, which corresponds to a lower threshold. The results of several tests conducted on multiple crop images at various growth stages showed that takes a value of (-8.67) and takes a value of 0.354, resulting in relatively favorable outcomes concerning image segmentation.
As illustrated in Figure 3B, the vertical projection method will be reused on centerline images to carry out the segmentation process mentioned above. Specifically, a horizontal line will be generated to intersect with the vertical projection curve, and the region enclosed by the intersection points is considered the final fitting area. The horizontal line shown in Figure 6B is given by Equation 18:
where represents the maximum y-value on the vertical projection curve of the centerline image. The value is optional, and since the origin of the coordinate system in OpenCV is in the top-left corner, the corresponding horizontal line y-value should be calculated starting from the origin.
After the aforementioned preprocessing steps, the resulting image was nearly devoid of noise or outliers. Based on the resulting image, this study utilized the advantages of the RANSAC algorithm (Fischler and Bolles, 1981) to select a group of representative points with higher density. Subsequently, the least squares method shown in Equation 21 was applied to fit these points, avoiding the fitting line biased towards the density center of the points. The fitting effect is improved, as shown in Supplementary Figure S2C. Furthermore, as depicted in Figure 6C, it was worth noting that not all the pixels are required for navigation line fitting. The inclusion of additional pixels leads to increased calculative complexity. As a result, feature points that can capture the main characteristics of the crop should be selected before the fitting process. The aforementioned sliding window method is an effective technique to extract feature points. As shown in Supplementary Figure S2A, the data has been significantly simplified while effectively preserving the primary characteristics of the crop. In this case, the sliding window has a width of 16 and a height of 32.
The RANSAC is a widely used robust estimation technique that is critical in various computer vision and machine learning applications. The RANSAC algorithm can be expressed as shown in Equation 19. Two crucial parameters of RANSAC include the distance threshold and the number of iterations. The distance threshold can be chosen based on either empirical observations or experiments, and this study sets it to 0.155. The number of iterations is given by Equation 20.
where represents the given set of points, represents the parameter set for the line to be fitted, is the distance from the point to the line, is the optimal parameter for the line, represents the size of the set and represents the distance threshold.
where represents the probability of the correct model, represents the probability of the correct model and which is set to 2 represents the minimum number of points required for each model estimation. To enhance model accuracy, is set to 0.99 in this study and takes a value of 2 in the initial iteration.
where represents the optimal inlier set.
A set of crop images at different growth stages was selected for rigorous testing and evaluation to assess the efficacy of the aforementioned image-processing techniques. To quantitatively analyze the effectiveness of the image processing results, this study employed a manual annotation method to obtain ideal reference lines for crop row navigation in different images, and the obtained reference lines were regarded as a criterion. The reference lines were subsequently juxtaposed with the fitting lines, and the accuracy of the system was evaluated by computing the angular discrepancies between the reference lines and the fitted lines. The yaw angle between the fitting line and the reference line is calculated by the following Equation 22:
where and represent the vectors of two lines, respectively. To enhance the analysis of angular data, the following mathematical metrics for evaluation are employed: MEA, RMSE, and MRE, which are defined below.
where and represent the MEA and the yaw angle between two lines of image, respectively, and represent, respectively, the angles between the fitting line and the manually annotated line with the horizontal line at the bottom of the image. is the total number of image samples.
The smaller the MEA is, the closer the fitting result is to the reference line. RMSE takes into account the size and distribution of the errors. MRE offers insights into the smaller the angle error relative to the reference line.
To facilitate effective user interaction and visual feedback, a human-computer interaction interface was designed and developed in this section, as illustrated in Supplementary Figure S3. This interaction interface contains resolution selection, processing, processing time and result display. This interface serves as a critical bridge between the algorithm and its real-world applications. Once finishing the resolution selection and processing, the result can be shown in the interaction interface.
This section involves carefully curating a subset of sugarcane seedling images from ‘Code 12’ to ‘Code 19’ to serve as a validation dataset for assessing the efficacy of the image processing techniques on sugarcane. The yaw angles between the fitting lines and the reference lines are illustrated in Figure 7.
Figure 7. Navigation line extraction results of sugarcane seedlings with different growth stages. (A) Code 12; (D) Code 14; (B, E) Code 16; (C, F) Code 19; (A–F) illustrate the yaw angles of sugarcane at different growth stages.
Table 1 illustrates the evaluation of three growth stages, revealing an initial upward trend followed by a subsequent decline in MEA, RMSE, and MRE. This pattern can be attributed to variations in branch distribution. Specifically, during stages of ‘Code 12-13’, sugarcane seedling foliage is small and less dispersed, resulting in minimal data fitting dispersion and relatively good symmetry. However, during stages of ‘Code 14-16’, the foliage becomes more extensive and dispersed, leading to increased data fitting dispersion and diminished symmetry. Notably, in stages of ‘Code 17-19’, the foliage is large and evenly distributed, resulting in a high degree of data fitting dispersion but maintaining good symmetry.
This section provides a comprehensive evaluation of the image processing techniques applied to corn and rice seedling images, primarily focusing on the growth stages corresponding to ‘Code 12-15’ and ‘Code 16-19’, which aids in assessing the algorithm generality and applicability across crop image scenarios. The yaw angles between the fitting lines and the reference lines are illustrated in Figures 8 and 9, respectively. The evaluation of corn and rice seedlings is denoted in Tables 2 and 3, respectively.
Figure 8. Navigation line extraction results of corn seedlings with different growth stages. (A, B, D, E) Code 12-15; (C, F) Code 16-19; (A-F) illustrate the yaw angles of corn at different growth stages.
Figure 9. Navigation line extraction results of rice seedlings with different growth stages. (A) Code 12; (D) Code 15; (B, E) Code 17; (C, F) Code 19; (A-F) illustrate the yaw angles of rice at different growth stages.
Table 2 presents the evaluation results for two growth stages of corn seedlings. In the growth stage of ‘Code 12-15’, the algorithms exhibit enhanced performance on corn seedlings, with lower MEA, RMSE, and MRE compared to crops in the growth stage of ‘Code 16-19’. The primary factors contributing to this phenomenon are the morphological differences in corn during these two growth stages and weeds. As the growth period progresses, both the dispersion of corn leaves and the presence of weeds in the field can, to some extent, impact the accuracy of navigation line extraction.
Contrary to corn, the algorithms demonstrate better performance in the growth stage of ‘Code 16-19’ of rice compared to the growth stage of ‘Code 12-15’. The uneven dispersion of leaves in rice during the early growth stage has been well documented and attributed to the long length of the leaves. As the plant grows and more leaves are produced, the dispersion becomes more even.
Seedling less ridges are prevalent in agricultural cultivation (see Figure 10), especially for multi-year crops. Such ridges result in the wastage of land, fertilizers, and other resources, ultimately impacting crop quality and yield. Achieving precise navigation in rows affected by seedling-less ridges ensures efficient execution of tasks such as replanting and fertilization application.
Figure 10. Navigation line extraction results of seedling less ridges. (A-F) illustrate instances of crop seedling absence within specific sections of a crop row. The white circle indicates the location of the seedling absence.
As shown in Figure 10, the navigation line can be extracted even in severe crop seedling absence in the crop rows. The MEA, RMSE, and MRE indicators in Table 4 also demonstrate that the extracted navigation line meets high precision requirements.
Based on the analysis in Section 2.2.3, the imaging method employed in this study identifies the baseline as the area with the highest pixel density along the vertical direction of the image, corresponding to the potential central crop row. By establishing the region of interest based on this baseline, the algorithm can dynamically classify crops belonging to the same row, unaffected by missing crops. The accurate extraction of the median crop row within the ROI enables the algorithm to effectively handle seedling absence. By comparing the distances between detected crop clusters and the baseline, the algorithm can robustly identify the correct row despite missing or sparse crop seedlings. This leads to improved accuracy and stability in scenarios with crop row discontinuities. These results demonstrate the centerline selection approach significantly enhances the robustness of crop row detection in challenging conditions.
The image processing techniques presented in this study demonstrate good generalization ability for corn and rice seedlings and perform well in extracting navigation lines in complex field environments, as shown in Figure 11. The evaluation of diverse in-field environments is denoted in Table 5.
Figure 11. Navigation line extraction results of different crop seedlings in the complex field environment. (A) Variations in illumination conditions: (Aa) Intense illumination; (Ab) Shadowed environment; (Ac) Water surface glare; (B) Decaying branches and fallen leaves: (Ba) Dried leaves; (Bb) Large clusters of dried branches; (Bc) Fallen rice leaves; (C) Variations in soil backgrounds: (Ca) Yellowish soil; (Cb) Dark-colored soil; (Cc) Pale-green soil; (D) Chaotic vegetation: (Da) Adjacent crops interconnection and occlusion; (Db) Weed and misaligned crop; (Dc) Duckweed and green algae.
As shown in Table 5, it can be inferred that the algorithms are robust in dealing with images derived from complex in-field environments, especially for those images regarding illumination and disordered vegetation.
Compared to branch-and-leaf and soil environments, the algorithm in this study demonstrates better robustness under varying illumination conditions and in chaotic vegetation environments. This can be attributed to the algorithm primary focus on green pixels in the image, while morphological operations based on NZPR effectively eliminate redundant pixel information. However, in branch-and-leaf and soil environments, factors such as fallen leaves and the greenish appearance of soil in paddy fields introduce a certain degree of disorder to the green pixel information, resulting in a slight decrease in algorithm performance.
Data augmentation plays a crucial role in improving the robustness and generalization capability of machine learning models, especially when addressing variations in environmental conditions. In this section, the Albumentations library was utilized to simulate real-world scenarios under low-light and rainy conditions. This evaluation aims to further validate the algorithm adaptability to special conditions and enhance its performance in practical applications.
As shown in Table 6, the navigation line extraction accuracy of traditional image processing algorithms exhibits significant differences under low-light and rainy conditions. According to the data, rainy conditions have a lesser impact on navigation line extraction, with lower MAE, RMSE, and MRE values for sugarcane, corn, and rice compared to low light conditions, indicating good algorithm robustness in rainy environments. In contrast, under low light conditions, the errors increase significantly, primarily due to insufficient lighting, which restricts the extraction of green features in the images. This is particularly evident in extreme low light environments, as illustrated in Figures 12Ac, Cc, Ec, where green features are substantially suppressed. Although rainy environments also involve some degree of low light, the overall lighting is typically better than in pure low-light conditions. Among the crops, corn shows the lowest error across both conditions, while rice exhibits relatively higher errors, especially in terms of RMSE and MRE, highlighting the challenges faced by the algorithm in these scenarios.
Figure 12. Navigation line extraction results of different crop seedlings in simulated environments. As shown in (Bb1), dim grayish raindrops are distributed across the image at a certain inclined angle. (A) Low light for sugarcane seedlings; (B) Rainy for sugarcane seedlings; (C) Low light for corn seedlings; (D) Rainy for corn seedlings; (E) Low light for rice seedlings; (F) Rainy for rice seedlings.
In high-resolution image processing, the increase in the number of pixels necessitates the processing of a larger number of data points, which significantly augments the computational load and processing time. To address this challenge, code and algorithm optimizations were implemented. Ten images of three crop seedlings at different growth stages were used to evaluate the average time consumption of various modules in the proposed algorithm. As shown in Figure 13, the processing time for a late-growth stage crop image with a resolution of 1920×1080 is largely dominated by the DBSCAN clustering algorithm. This is attributed to the high complexity of the neighbor search process, where processing time escalates markedly with increasing resolution. Based on Figure 13, it is evident that the processing time of the optimized DBSCAN algorithm has been significantly reduced. At a resolution of 640×480, the processing time was nearly 30%. Meanwhile, at a resolution of 1920×1080, the processing time was reduced by almost 94%. However, it is worth noting that the processing time of the optimized DBSCAN algorithm still accounts for the majority of the total processing time.
Figure 13. Comparison of average processing times before and after optimizations for different main algorithm modules (processing time exceeding 1 ms). The numbers 1-7 correspond to the following algorithm modules: ExG, OTSU, NZPR-Morphological Operation, Eight-Connectivity, first sliding window, KD-Tree accelerated DBSCAN, second sliding window. (A) Resolution of 640×480; (B) Resolution of 1920×1080.
To address the issue of DBSCAN, this study has implemented a KD-Tree to accelerate the neighbor search and utilized OpenMP for optimizing parallelization. Crop images with two resolutions (640×480 and 1920×1080) were utilized to test the processing time of the algorithm proposed in this paper, with 20 images of crops under different environmental conditions selected for each growth stage, as shown in Figure 14. The average processing times for the crops at these two resolutions are presented in Table 7.
Figure 14. Processing times of crop images at the two resolutions. (A) Resolution of 640×480 before optimization; (B) Resolution of 640×480 after optimization; (C) Resolution of 1920×1080 before optimization; (D) Resolution of 1920×1080 after optimization; (a) Sugarcane; (b) Rice; (c) Corn.
It is observed that processing times for crops in the early growth stages (as exemplified by instance No. 1 in Figure 14Ab and No. 20 in Figure 14Bb) could be comparable to or exceed those recorded in later stages (as exemplified by instance No. 15 in Figures 14Ab, Bb). This discrepancy is likely attributable to varying field conditions; for instance, images from earlier growth stages may contain substantial areas of weeds.
A comparison between Figures 14A, B, as well as Figures 14C, D, reveals a significant reduction in processing time after optimization, with the improvement being particularly pronounced in high-resolution image processing. Although there is some variation in processing times across different crop images, the optimized algorithm exhibits much greater stability, indicating that the optimization not only reduces processing time but also enhances the stability of performance. However, even after optimization, the processing time for high-resolution images remains significantly longer than that for low-resolution images, highlighting the continued computational challenges posed by high-resolution data.
As shown in Table 7, the evaluation suggests that a resolution of 640×480 offers an optimal balance for practical applications. Comparing the average processing times before and after optimization, for images with a resolution of 640×480, the average processing time decreased by approximately 25% to 30% after optimization. The optimization effect is even more pronounced for images with a resolution of 1920×1080, where the processing time was reduced by approximately 80% to 86%. The processing time was significantly reduced across all resolutions and crop types, indicating that the optimization methods employed, such as DBSCAN accelerated using KD-Tree and OpenMP parallelization, are effective in practical applications. As a result, the selection of resolution is decided by personalized needs.
The proposed image-processing techniques exhibit effective performance in navigation line extraction for sugarcane seedling plantation rows, accompanied by a desired generalization capability when applied to different crops and diverse, complex in-field environments. This study bridges a gap that previous studies still need to address.
This study proposes a crop row navigation line extraction algorithm designed based on the principle of prioritizing green pixels. The algorithm utilizes traditional image processing techniques, starting with a preprocessing stage that enhances green features and effectively removes noise through grayscale transformation and OTSU segmentation. Morphological operations guided by the NZPR significantly reduce unnecessary pixels in the image, isolating the main stem regions of the crop rows. Noise caused by the morphological operations is further removed using an eight-connected algorithm, which also refines the stem regions. During this step, vertical projection is applied to detect curve peaks (baseline), providing a foundation for clustering the central crop row. Optimal region detection is performed using DBSCAN enhanced with KD-tree acceleration and parallelization techniques, enabling efficient pixel clustering while reducing computational time. The baseline detection process helps identify potential target crop clusters, and the trapezoidal ROI ensures the algorithm focuses on relevant areas, even in cases of crop absence, while excluding irrelevant clusters. Sigmoid thresholding further segments the target crop rows, producing the optimal region for subsequent processes. Finally, the RANSAC algorithm combined with the least squares method is employed to fit the data points. This ensures robust accuracy even in the presence of outliers, successfully extracting the navigation line. Image preprocessing is a critical step in the navigation line fitting process. During this stage, the number of extracted pixels directly impacts the algorithm performance: too few pixels may fail to adequately represent the characteristics of the crop row, while too many pixels could significantly increase the computational burden in subsequent steps. Therefore, ensuring the extracted image reflects the crop row main features is a key objective in the preprocessing stage.
To evaluate the adaptability of this step under different environmental conditions, this study specifically selected low-light and rainy conditions as experimental scenarios, given their broader and more pronounced impact on image quality compared to other factors. As shown in Figure 15B, the total number of pixels extracted during preprocessing is highest under normal conditions, followed by low-light conditions, and lowest under rainy conditions. This indicates that the algorithm is affected to some extent under extreme environments. However, the extracted image pixels remain effective in representing the main features of the crop rows. Moreover, the navigation lines obtained under the three different conditions are nearly identical, demonstrating the proposed algorithm robustness and adaptability in extreme environments.
Figure 15. Illustration of image preprocessing under extreme environments. (A) Original images: (Aa) Normal; (Ab) Low light; (Ac) Rainy; (B) Images after preprocessing; (C) Results of navigation line.
In this study, 640×480 and 1920×1080 are adopted as the target resolutions for image processing. These two resolutions reflect the consideration of mainstream resolutions in current visual devices and technology trends (Lin et al., 2020; He et al., 2022). The 640×480 resolution is widely used in real-time visual processing systems since that resolution requires reduced computational power, enable efficient algorithm operation within constrained computational resources. Conversely, 1920×1080, as a representative of higher definition resolutions, provides richer details and higher image quality, suitable for applications demanding higher detail accuracy, such as precise image analysis and advanced visual recognition tasks (Yang et al., 2020). By conducting experiments at these two resolutions, the adaptability of our proposed algorithm to different application demands can be validated. When applying our proposed algorithm to perform diverse field management tasks at the seedling stage such as weeding, spraying pesticides, and root-zone soil backfill, an optimal resolution can be selected through our developed human-machine interface in response to cross growth stage of crop seedlings.
Statistical analysis reveals that the algorithm average runtime is 0.31 seconds for images with a resolution of 640×480 and 0.51 seconds for images with a resolution of 1920×1080. In practical applications, agricultural robots performing field management tasks such as weeding, spraying pesticides, fertilizer application, and root-zone soil backfill typically travel at 0.6~1.5 meters per second. This indicates that the algorithm proposed in this study addresses the demands for real-time application.
The excellent adaptability of the algorithm to crops at different growth stages represents a significant breakthrough in this study. The average values of the growth stages of ‘Code 12-19’ from the evaluation results of Tables 1–3 show that MEA is 2.01°, RMSE is 2.95°, and MRE is 2.37%, which has demonstrated commendable performance in the case of cross-growth-stage crops. Additionally, by comparing the experimental results of different crops, it can be observed that the algorithms perform better on corn and rice seedlings, which have relatively concentrated and evenly distributed branches and leaves, than on sugarcane seedlings, which have unevenly dispersed branches and leaves.
An essential aspect of evaluating the performance and robustness of the proposed image processing techniques is the adaptability to complex field environments. Complex field environments, as depicted in Figure 11, can pose various challenges for the algorithms, such as varying illumination (Figure 11A), fallen leaves (Figure 11Ba), dry branches (Figure 11Bb), soil background (Figure 11C), similar colors (Figure 11Db) and occlusion (Figure 11Da), dry fields (the first and second columns in Figure 11) and paddy land (the third column of Figure 11), different crop row spacing, etc. These factors can affect the accuracy and efficiency of the algorithms in detecting and segmenting crop plants from the images. Therefore, the algorithm adaptability to different complex field environments will be discussed in this section.
On the basis of the evaluation results of environments presented in Tables 4 and 5, it can be inferred that MEA has an average value of 1.0018°, while RMSE has an average value of 1.446°. Moreover, MRE has an average value of 1.126%. Among the environments presented in Table 5, the algorithms perform the best in chaotic vegetation, indicating that they can effectively handle situations where the branches and leaves of crops overlap and occlude with each other, while also accounting for the presence of offset crops and weeds. This remarkable performance can be mainly attributed to the proposed NZPR concept and the mapping relationship between NZPR and morphological operations. It is noteworthy that the common phenomenon of seedling less ridges in agricultural production can be successfully clustered into crop rows, thanks to the centerline selection method proposed in this study.
The inter-row spacing of the crops tested in this study was 1 meter for sugarcane seedlings (Figure 7), 0.5 meters for corn seedlings (Figure 8), and 0.3 meters for rice seedlings (Figure 9). Upon analyzing the evaluation results of three crops, it becomes evident that the algorithms can effectively facilitate precise navigation for crops with inter-row spacing variability. As illustrated in section 2.2.3 and section 2.3.2, the utilization of vertical projection can successfully detect the baseline, which serves as the reference for centerline selection; crop clusters located in the centerline area can be determined by applying the distance threshold method, thus ensuring adaptability to inter-row spacing variability.
Arid land and paddy fields represent two major environments for crop cultivation, bearing considerable influence on agricultural production. For arid land crops, represented by corn, the average values calculated from Table 2 present that MEA is 1.39°, RMSE is 1.844°, and MRE is 1.665%. For paddy field crops, represented by rice, the average values calculated from Table 3 present that MEA is 2.3195°, RMSE is 3.4°, and MRE is 2.6145%. It is easily noticeable that the algorithm exhibits better adaptability to corn seedlings than rice seedlings. Apart from the influence caused by the inherent morphological differences between these two crops, factors such as reflections and glare in the paddy field environment can also interfere with the extraction of navigation lines. However, even in the case of rice seedlings, where the performance is relatively less favorable, there is a slight improvement in terms of MEA when compared to similar studies in the same category (Fu et al., 2023; Wang A. et al., 2021).
In practical agricultural operations, it is inevitable to encounter extreme conditions such as low light, nighttime, and rainy weather. Therefore, simulating these extreme environments, as shown in Figure 12, through image augmentation is highly beneficial for further validating the generalization capability of the proposed algorithm. According to the results in Table 6, the average values of MEA, RMSE, and MRE for the three crops under low light and rainy conditions are 1.41°, 2.67°, and 1.56%, respectively. When compared with the average values obtained under various real-world conditions in Tables 4 and 5 (MEA: 1.0018°, RMSE: 1.446°, MRE: 1.126%), it is evident that low-light and rainy conditions present greater challenges to the robustness and generalization capability of the proposed algorithm. This discrepancy can be attributed to the fact that lighting and rain affect the image globally, leading to diminished visibility of green features under such conditions. Since the algorithm proposed in this paper heavily relies on green information for subsequent processing, the decrease of these features in low-light and rainy environments directly impacts its performance. This highlights that future research could explore the use of sensors to acquire multidimensional data of crops, addressing the limitations of relying solely on image-based information.
On the basis of our constrained image dataset, to accelerate the training process and improve the model accuracy and generalization, we use transfer learning to obtain a model that can identify rice, sugarcane and corn plants. A YOLOv8n pre-trained weights was used in the process of transfer learning. The software environment comprised Python version 3.9, PyTorch version 2.4.1, and CUDA version 12.1. The model was trained with a batch size of 2 and 300 epochs, while all other hyperparameters were kept at their default settings. In this study, our datasets consisting three types of crop seedlings under various environmental conditions were also used for training. The primary stem regions of the seedlings were annotated, aligning with the approach of the proposed traditional image processing algorithm, which extracts the main stem regions for navigation line fitting. The workflow is illustrated in Figure 16. The process begins with training the YOLOv8n model on the dataset. Once trained, the model processes the test set to generate bounding boxes around the crop regions. The center points of these bounding boxes are extracted and clustered using the DBSCAN algorithm to identify the primary crop row. The clustered center points are then used in an Orthogonal Distance Regression (ODR) model to fit a navigation line.
Figure 16. The workflow for navigation line extraction combining transfer learning and image processing methods (named as CTLIP).
Supplementary Table S1 illustrates the evaluation results of navigation line extraction using the YOLOv8n model for three crops: sugarcane, corn, and rice. Among the crops, rice achieved the best overall performance, with the lowest MAE (0.833°), RMSE (1.225°), and MRE (0.971%), indicating that the model could effectively detect and extract navigation lines in rice fields. This performance can be attributed to the relatively distinct stem regions of rice seedlings, which are easier to identify, and the minimal occlusion between individual stems. In contrast, sugarcane and corn showed higher error values, with corn having the highest MRE (2.386%) and sugarcane having the highest RMSE (2.934°). The inferior performance of sugarcane and corn can be attributed to their severe occlusion in the later growth stages, where stem regions become nearly indistinguishable. These occluded areas significantly increase the occurrence of missed detections (shown in the white circled area of Figure 17), thereby affecting the accuracy of navigation line extraction. Additionally, the performance was observed to degrade under extreme low-light conditions, where green features are heavily suppressed. This leads to higher error rates and occasional missed detections. These results will serve as the basis for a comparative analysis with the algorithm proposed in this study, which will be detailed in the subsequent discussion section. The aim is to evaluate the advantages and limitations of deep learning in handling various conditions relative to the approach proposed in this study.
Figure 17. Navigation line extraction results of CTLIP. (A) Sugarcane seedlings; (B) Corn seedlings; (C) Rice seedlings.
To further present our study, in combination with evaluation metrics used in this study, a comparison analysis was done between our study with existing similar research and the CTLIP method, as illustrated in Supplementary Table S2.
As shown in Supplementary Table S2, the proposed navigation line extraction algorithm outperforms the traditional image processing method by Fu et al. (2023) in terms of MEA, demonstrating a notable reduction in error. Moreover, the RMSE achieved by our algorithm is 2.47°, further highlighting its accuracy across multiple crops, growth stages and diverse field environments, including sugarcane and corn, as opposed to the single crop tested in the comparative study. Additionally, the inclusion of MRE in our results presents a more comprehensive evaluation. This demonstrates the enhanced generalization and robustness of our algorithm in terms of traditional image processing. The comparison between CTLIP and the proposed algorithm indicates that both methods exhibit good performance in crop row detection. Our method achieved better results, reflecting its high precision and effective error control. This highlights the capability of traditional image processing methods to leverage prior knowledge and structured algorithms for accurate navigation line extraction in simpler, well-defined scenarios. Notably, CTLIP performed particularly well in scenarios where the main stems of crops were visible with minimal occlusion, such as in rice seedlings (lowest MEA of 0.833° in Table 7). Both methods outperformed the Fu et al. (2023) approach, emphasizing the adaptability across multiple crops (sugarcane, corn, and rice) and environmental conditions. The findings suggest that integrating the strengths of traditional methods and deep learning could further enhance crop row detection accuracy, particularly under diverse field conditions.
In this study, traditional image processing methods and deep learning-based object detection methods exhibit distinctly different challenges in navigation line extraction. For traditional image processing, a significant source of error lies in the clustering process, particularly when crop rows appear closer together at the top of the image. This issue is exacerbated in cases where the inter-row spacing is narrower, as shown in Figure 18A, C with rice seedlings. The clustering algorithm may mistakenly group crops from adjacent rows into the central crop row, especially in the top regions of the image, leading to inaccuracies in navigation line fitting. A practical solution to this problem is to adjust the field of view of the camera. Since real-time navigation typically does not require distant views, narrowing the field of view can reduce the apparent convergence of crop rows at the top of the image, thereby mitigating clustering errors.
Figure 18. Illustrations of error samples. (A, C) Narrower interrow spacing; (B, D) Missing detection.
In contrast, deep learning-based object detection errors primarily stem from missed detections. As illustrated in Figures 18B, D, missed detections result in the loss of critical information from both the top and bottom regions of the crop rows in the image. When fitting the navigation line, the algorithm can only rely on a limited number of bounding boxes, significantly impacting navigation line extraction accuracy. In extreme cases, such as when only one bounding box remains for the central crop row, it becomes impossible to perform navigation line fitting, which is a critical failure for the algorithm. The reasons for these missed detections include the inability to extract features of the main stem. This issue can arise due to severe occlusion among crops, where overlapping leaves and stems obscure the main stem from the camera view. Additionally, in low-light conditions or during inclement weather, the features of the crops become less distinct in the images. This lack of clarity hampers the deep learning model ability to accurately detect and identify the main stems, leading to increased instances of missed detections. To address this issue, future research will focus on collecting more crop datasets under varying environmental conditions and using more efficient and suitable modules in the deep learning model.
This comparative analysis highlights that, in this study, traditional methods struggle with row spacing and clustering errors, and deep learning methods are vulnerable to detection omissions, especially in adverse environmental or crop conditions. Understanding these failure modes is essential for improving the robustness of navigation line extraction in practical agricultural scenarios.
Deep learning models have achieved impressive progress in image recognition but still have some shortcomings and challenges. One such challenge is that most existing models are trained for specific datasets and are usually unable to adapt to different scenarios, thus lacking generalization ability. Additionally, the scale and diversity of the required datasets are demanding. Furthermore, the high computational cost and latency associated with deep learning may not be conducive to some practical applications.
In extracting navigation lines, deep learning-based object detection is well-suited for scenarios where there is a clear differentiation between the growth stages, minimal crop occlusion, and easy plant identification. However, these ideal field conditions are rare in most cases. As a result, the algorithms proposed in this study fall under classical image processing techniques, characterized by their cost-effectiveness, minimal data requirements, ease of integration with edge devices, and suitability for deployment on resource-limited embedded systems or mobile devices. In addition, classical image processing techniques are more interpretable and explainable, as they rely on well-defined mathematical models and algorithms rather than black-box neural networks. This is particularly desirable for researchers, as it helps to understand the fundamental logic governing variations in phenomena, thereby enabling adequate explanation and management. Notably, deep learning-based object detection outperforms traditional image processing methods in the presence of weeds within crop rows. This advantage arises because deep learning models are capable of learning complex features and distinguishing between crops and weeds based on extensive training datasets, whereas traditional methods often rely on predefined rules and thresholds, which can be easily disrupted by the presence of similar or overlapping features such as weeds.
Overall, the presented study highlights the generalization of navigation line extraction at different crop seedling stages. Despite advancements in classical image processing techniques, there are limitations in dealing with complex image processing tasks such as complex seedling interaction between adjacent planting rows, intense weeds between plantations.
In this study, a pixel-wise navigation line extraction method of cross-growth-stage seedlings was verified, showing a promising solution for in-depth intelligent agricultural machinery application, especially for the field management task practices such as stumping, weeding, spraying pesticides, and root-zone soil backfill. For future studies, the proposed navigation line extraction method can be integrated into agricultural machinery, make it possible to change the current field management task practices from mechanization to intelligence, solving the problem of population ageing and labor costs.
From the results presented in this study, the following conclusions could be drawn:
1. The present study successfully achieved near-noiseless binary images through pre-processing. It identified appropriate fitting regions by detecting crop rows. Furthermore, it obtained precise navigation lines through linear fitting, regardless of growth stages, diverse field environments, or ridges of seedling absence.
2. The performance of the proposed algorithms was comprehensively evaluated by three indicators: MEA, RMSE, and MRE. For different crops under different growth stages or environmental conditions, MEA and RMSE are basically in the range of 1° to 4°, while MRE is in the range of 1% to 4%, which indicates that the navigation line and the reference line were closely aligned. The distribution of errors between them was found to be relatively uniform. For the 640×480 resolution, the algorithm average processing time is 310 ms, demonstrating sufficient real-time performance to meet the navigation and operational requirements of agricultural robots. For the 1920×1080 resolution, the average processing time is 510 ms, which, despite being longer, is still adequate for practical applications, ensuring that the algorithm maintains real-time functionality in field management tasks.
3. The mathematical relationship between opening operation and NZPR can create clear and distinct boundaries between crop rows in different growth stages. This enables the connected component filter to significantly suppress irrelevant information in the image. Moreover, the fitting region that best represents the central part of the crop can be obtained by applying the NZPR adaptive sigmoid function to segment the vertical projection curves of different centerlines. This ensures the accuracy of the centerline fitting.
In conclusion, this study proposed an image processing algorithm for navigation line extraction of cross-growth-stage seedlings in sugarcane, corn and rice. Its efficiency and generalization capabilities were verified in complex field environments.
Publicly available datasets were analyzed in this study. This data can be found here: https://github.com/Mr-lxd/fieldNavigationProject, https://doi.org/10.6084/m9.figshare.24188097.v5.
HL: Conceptualization, Funding acquisition, Project administration, Resources, Supervision, Writing – original draft, Writing – review & editing. XL: Data curation, Investigation, Methodology, Software, Resource, Validation, Visualization, Writing – original draft, Writing – review & editing. YM: Resources, Writing – original draft. DH: Writing – review & editing. TW: Funding acquisition, Writing – review & editing.
The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This work was supported by the Youth Science Foundation of Guangxi Province (grant No. 2024JJB160121), Young Talents Funding Research Projects of Guangxi Province (grant No. ZX02080030124006), Guangxi Key Laboratory of Manufacturing System and Advanced Manufacturing Technology (grant No. 22-35-4-S306), the Sugarcane Whole-process Mechanization Scientific Research Base of the Ministry of Agriculture and Rural Affairs, and the National Sugar Industry Technology System (grant No. CARS-170402).
The authors would like to thank the editor and the reviewers for their constructive comments and helpful suggestions.
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
The reviewer YL declared a shared affiliation with the author TW to the handling editor at the time of review.
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpls.2024.1499896/full#supplementary-material
Supplementary Figure 1 | Image graying process by applying ExG. (A) Original image; (B) Grayscale image.
Supplementary Figure 2 | Navigation line fitting. (A) Crop pixels after applying the sliding window method; (B) The navigation line is obtained after applying RANSAC and the least squares method.
Supplementary Figure 3 | Interaction interface for field navigation at crop seedling stage.
Bai, Y., Zhang, B., Xu, N., Zhou, J., Shi, J., Diao, Z. (2023). Vision-based navigation and guidance for agricultural autonomous vehicles and robots: A review. Comput. Electron. Agric. 205, 107584. doi: 10.1016/j.compag.2022.107584
Blok, P. M., van Boheemen, K., van Evert, F. K., IJsselmuiden, J., Kim, G. H. (2019). Robot navigation in orchards with localization based on Particle filter and Kalman filter. Comput. Electron. Agric. 157, 261–269. doi: 10.1016/j.compag.2018.12.046
Chen, J., Qiang, H., Wu, J., Xu, G., Wang, Z., Liu, X. (2020). Extracting the navigation path of a tomato-cucumber greenhouse robot based on a median point Hough transform. Comput. Electron. Agric. 174, Article 105472. doi: 10.1016/j.compag.2020.105472
Chen, J., Qiang, H., Wu, J., Xu, G., Wang, Z. (2021). Navigation path extraction for greenhouse cucumber-picking robots using the prediction-point Hough transform. Comput. Electron. Agric. 180, Article 105911. doi: 10.1016/j.compag.2020.105911
Chen, X., Wang, S., Zhang, B., Luo, L. (2018). Multi-feature fusion tree trunk detection and orchard mobile robot localization using camera/ultrasonic sensors. Comput. Electron. Agric. 147, 91–108. doi: 10.1016/j.compag.2018.02.009
Chen, Z., Li, W., Zhang, W., Li, Y., Li, M., Li, H. (2019). Vegetable crop row extraction method based on accumulation threshold of Hough Transformation. Trans. Chin. Soc Agric. Eng. 35, 314–322. doi: 10.11975/j.issn.1002-6819.2019.22.037
Diao, Z., Guo, P., Zhang, B., Zhang, D., Yan, J., He, Z., et al. (2023). Navigation line extraction algorithm for corn spraying robot based on improved YOLOv8s network. Comput. Electron. Agric. 212, 108049. doi: 10.1016/j.compag.2023.108049
Diao, Z., Ma, S., Zhang, D., Zhang, J., Guo, P., He, Z., et al. (2024). Algorithm for corn crop row recognition during different growth stages based on ST-YOLOv8s network. Agron 14, 1466. doi: 10.3390/agronomy14071466
Diao, Z., Zhao, M., Song, Y., Wu, B., Wu, Y., Qian, X., et al. (2015). Crop line recognition algorithm and realization in precision pesticide system based on machine vision. Trans. Chin. Soc Agric. Eng. 31, 47–52. doi: 10.3969/j.issn.1002-6819.2015.07.007
Ester, M., Kriegel, H. P., Sander, J., Xu, X. (1996). “A density-based algorithm for discovering clusters in large spatial databases with noise,” in Proceedings of the Second International Conference on Knowledge Discovery and Data Mining (Portland, Oregon: AAAI Press), 226–231. doi: 10.5555/3001460.3001507
FAO (2022). FAO’s Statistical Yearbook for 2022 goes live. Available online at: https://www.fao.org/newsroom/detail/fao-s-statistical-yearbook-for-2022-goes-live/en (Accessed August 18, 2023).
Fischler, M. A., Bolles, R. C. (1981). Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 24, 381–395. doi: 10.1145/358669.358692
Fu, D., Jiang, Q., Qi, L., Xing, H., Chen, Z., Yang, X. (2023). Detection of the centerline of rice seedling belts based on region growth sequential clustering-RANSAC. Trans. Chin. Soc Agric. Eng. 39, 47–57. doi: 10.11975/j.issn.1002-6819.202210106
Gong, J., Sun, K., Zhang, Y., Lan, Y. (2022). Extracting navigation line for rhizome location using gradient descent and corner detection. Trans. Chin. Soc Agric. Eng. 38, 177–183. doi: 10.11975/j.issn.1002-6819.2022.13.020
Gong, J., Wang, X., Zhang, Y., Lan, Y. (2020). Extraction method of corn rhizome navigation lines based on edge detection and area localization. Trans. Chin. Soc Agric. Mach. 51, 26–33. doi: 10.6041/j.issn.1000-1298.2020.10.004
He, C., Chen, Q., Miao, Z., Li, N., Sun, T. (2021). “Extracting the navigation path of an agricultural plant protection robot based on machine vision,” in 2021 40th Chinese Control Conference (Shanghai, China: CCC), 3576–3581. doi: 10.23919/CCC52363.2021.9549671
He, R., Luo, X., Zhang, Z., Zhang, W., Jiang, C., Yuan, B. (2022). Identification method of rice seedlings rows based on Gaussian heatmap. Agriculture 12, 1736. doi: 10.3390/agriculture12101736
He, J., Meng, Q., Zhang, M., Qiu, R., Xiang, M., Du, S. (2014). Crop baseline extraction method for off-road vehicle based on boundary detection and scan-filter. Trans. Chin. Soc Agric. Mach. 45, 1. doi: 10.6041/j.issn.1000-1298.2014.S0.043
Jiang, G., Wang, Z., Liu, H. (2015). Automatic detection of crop rows based on multi-ROIs. Expert. Syst. Appl. 42, 2429–2441. doi: 10.1016/j.eswa.2014.10.033
Jiang, G., Yang, X., Wang, Z., Liu, H. (2017). Crop rows detection based on image characteristic point and particle swarm optimization-clustering algorithm. Trans. Chin. Soc Agric. Eng. 33, 165–170. doi: 10.11975/j.issn.1002-6819.2017.11.021
Li, D., Li, B., Kang, S., Feng, H., Long, S., Wang, J. (2023). E2CropDet: An efficient end-to-end solution to crop row detection. Expert. Syst. Appl. 227, 120345. doi: 10.1016/j.eswa.2023.120345
Li, X., Su, J., Yue, Z., Duan, F. (2022a). Adaptive multi-ROI agricultural robot navigation line extraction based on image semantic segmentation. Sensors 22, 7707. doi: 10.3390/s22207707
Li, X., Su, J., Yue, Z., Wang, S., Duan, F., Hua, J. (2022c). “Vision-based navigation line extraction by combining crop row detection and RANSAC algorithm,” in 2022 IEEE International Conference on Mechatronics and Automation (Guilin, Guangxi, China: ICMA), 1097–1102. doi: 10.1109/ICMA54519.2022.9856296
Li, X., Su, J., Yue, Z., Wang, S., Zhou, H. (2022b). Extracting navigation line to detect the maize seedling line using median-point Hough transform. Trans. Chin. Soc Agric. Eng. 38, 167–174. doi: 10.11975/j.issn.1002-6819.2022.05.020
Li, M., Zhang, M., Meng, Q., Liu, G. (2013). Rapid detection of navigation baseline for farm machinery based on scan-filter algorithm. Trans. Chin. Soc Agric. Eng. 29, 41–47. doi: 10.3969/j.issn.1002-6819.2013.01.006
Li, X., Zhao, W., Zhao, L. (2021). Extraction algorithm of the center line of maize row in case of plants lacking. Trans. Chin. Soc Agric. Eng. 37, 203–210. doi: 10.11975/j.issn.1002-6819.2021.18.024
Liao, J., Wang, Y., Yin, J., Zhang, S., Liu, L., Zhu, D. (2019). Detection of seedling row centerlines based on sub-regional feature points clustering. Trans. Chin. Soc Agric. Mach. 50, 34–41. doi: 10.6041/j.issn.1000-1298.2019.11.004
Lin, S., Jiang, Y., Chen, X., Biswas, A., Li, S., Yuan, Z., et al. (2020). Automatic detection of plant rows for a transplanter in paddy field using faster r-cnn. Ieee. Access. 8, 147231–147240. doi: 10.1109/ACCESS.2020.3015891
Liu, X., Qi, J., Zhang, W., Bao, Z., Wang, K., Li, N. (2023). Recognition method of maize crop rows at the seedling stage based on MS-ERFNet model. Comput. Electron. Agric. 211, 107964. doi: 10.1016/j.compag.2023.107964
Liu, R., Yandun, F., Kantor, G. (2024). LiDAR-Based Crop Row Detection Algorithm for Over-Canopy Autonomous Navigation in Agriculture Fields. Available online at: https://arxiv.org/abs/2403.17774 (Accessed November 11, 2024).
Liu, Z., Zhang, Z., Luo, X., Wang, H., Huang, P., Zhang, J. (2018). Design of automatic navigation operation system for Lovol ZP9500 high clearance boom sprayer based on GNSS. Trans. Chin. Soc Agric. Eng. 34, 15–21. doi: 10.11975/j.issn.1002-6819.2018.01.003
Luo, X., Zhang, Z., Zhao, Z., Chen, B., Hu, L., Wu, X. (2009). Design of DGPS navigation control system for Dongfanghong X-804 tractor. Trans. Chin. Soc Agric. Eng. 25, 139–145. doi: 10.3969/j.issn.1002-6819.2009.11.025
Ma, Z., Yin, C., Du, X., Zhao, L., Lin, L., Zhang, G., et al. (2022). Rice row tracking control of crawler tractor based on the satellite and visual integrated navigation. Comput. Electron. Agric. 197, 106935. doi: 10.1016/j.compag.2022.106935
Meng, Q., Liu, G., Zhang, M., Si, Y., Li, M. (2013). Crop rows detection based on constraint of linear correlation coefficient. Trans. Chin. Soc Agric. Mach. 44, 216–223. doi: 10.6041/j.issn.1000-1298.2013.S1.039
Meng, Q., Zhang, M., Qiu, R., He, J., Si, Y., Liu, G. (2014). Navigation line detection for farm machinery based on improved genetic algorithm. Trans. Chin. Soc Agric. Mach. 45, 39–46. doi: 10.6041/j.issn.1000-1298.2014.10.007
Meng, Q., Zhang, M., Yang, G., Qiu, R., Xiang, M. (2016). Guidance line recognition of agricultural machinery based on particle swarm optimization under natural illumination. Trans. Chin. Soc Agric. Mach. 47, 11–20. doi: 10.6041/j.issn.1000-1298.2016.06.002
Otsu, N. (1979). A threshold selection method from gray-level histograms. IEEE Trans. SMC 9, 62–66. doi: 10.1109/TSMC.1979.4310076
Rao, H., Ji, C. (2007). Crop-row detection using Hough transform based on connected component labeling. Trans. Chin. Soc Agric. Eng. 23, 146–150. Available online at: http://www.tcsae.org/en/article/id/20070330
Rao, X., Zhu, Y., Zhang, Y., Yang, H., Zhang, X., Lin, Y., et al. (2021). Navigation path recognition between crop ridges based on semantic segmentation. Trans. Chin. Soc Agric. Eng. 37, 179–186. doi: 10.11975/j.issn.1002-6819.2021.20.020
Ruan, Z., Chang, P., Cui, S., Luo, J., Gao, R., Su, Z. (2023). A precise crop row detection algorithm in complex farmland for unmanned agricultural machines. Biosyst. Eng. 232, 1–12. doi: 10.1016/j.biosystemseng.2023.06.010
Shi, J., Bai, Y., Diao, Z., Zhou, J., Yao, X., Zhang, B. (2023). Row detection based navigation and guidance for agricultural robots and autonomous vehicles in row-crop fields: Methods and applications. Agronomy 13, 1780. doi: 10.3390/agronomy13071780
Si, Y., Jiang, G., Liu, G., Gao, D., Liu, Z. (2010). Early-stage crop rows detection based on least square method. Trans. Chin. Soc Agric. Mach. 41, 163–185. doi: 10.3969/j.issn.1000-1298.2010.07.034
Song, Y., Xu, F., Yao, Q., Liu, J., Yang, S. (2022). Navigation algorithm based on semantic segmentation in wheat fields using an RGB-D camera. Inform. Process. Agric. 10, 475–490. doi: 10.1016/j.inpa.2022.05.002
Wang, X., Han, X., Mao, H., Liu, F. (2012). Navigation line detection of tomato ridges in greenhouse based on the least square method. Trans. Chin. Soc Agric. Mach. 43, 161–166. doi: 10.6041/j.issn.1000-1298.2012.06.029
Wang, W., Wang, J., Feng, C., Zhao, T., Shen, L., Feng, X., et al. (2023). Establishment of an efficient transgenic selection system and its utilization in Saccharum officinarum. Trop. Plants. 2, 11. doi: 10.48130/TP-2023-0011
Wang, S., Yu, S., Zhang, W., Wang, X. (2020). Detection of rice seedling rows based on Hough transform of feature point neighborhood. Trans. Chin. Soc Agric. Mach. 51, 18–25. doi: 10.6041/j.issn.1000-1298.2020.10.003
Wang, S., Yu, S., Zhang, W., Wang, X., Li, J. (2023a). The identification of straight-curved rice seedling rows for automatic row avoidance and weeding system. Biosyst. Eng. 233, 47–62. doi: 10.1016/j.biosystemseng.2023.07.003
Wang, S., Yu, S., Zhang, W., Wang, X., Li, J. (2023b). The seedling line extraction of automatic weeding machinery in paddy field. Comput. Electron. Agric. 205, 107648. doi: 10.1016/j.compag.2023.107648
Wang, A., Zhang, M., Liu, Q., Wang, L., Wei, X. (2021). Seedling crop row extraction method based on regional growth and mean shift clustering. Trans. Chin. Soc Agric. Eng. 37, 202–210. doi: 10.11975/j.issn.1002-6819.2021.19.023
Wang, S., Zhang, W., Wang, X., Yu, S. (2021). Recognition of rice seedling rows based on row vector grid classification. Comput. Electron. Agric. 190, 106454. doi: 10.1016/j.compag.2021.106454
Woebbecke, D. M., Meyer, G. E., Von Bargen, K., Mortensen, D. A. (1995). Color indices for weed identification under various soil, residue, and lighting conditions. Trans. ASAE 38, 259–269. doi: 10.13031/2013.27838
Yang, Y., Ma, Q., Chen, Z., Wen, X., Zhang, G., Zhang, T., et al. (2022). Real-time extraction of the navigation lines between sugarcane ridges using LiDAR. Trans. Chin. Soc Agric. Eng. 38, 178–185. doi: 10.11975/j.issn.1002-6819.2022.04.021
Yang, Y., Zhang, B., Zha, J., Wen, X., Chen, L., Zhang, T., et al. (2020). Real-time extraction of navigation line between corn rows. Trans. Chin. Soc Agric. Eng. 36, 162–171. doi: 10.11975/j.issn.1002-6819.2020.12.020
Yang, Y., Zhou, Y., Yue, X., Zhang, G., Wen, X., Ma, B., et al. (2023). Real-time detection of crop rows in maize fields based on autonomous extraction of ROI. Expert. Syst. Appl. 213, Article 118826. doi: 10.1016/j.eswa.2022.118826
Yu, Y., Bao, Y., Wang, J., Chu, H., Zhao, N., He, Y., et al. (2021). Crop row segmentation and detection in paddy fields based on treble-classification Otsu and double-dimensional clustering method. Remote Sens. Basel 13, 901. doi: 10.3390/rs13050901
Zadoks, J. C., Chang, T. T., Konzak, C. F. (1974). A decimal code for the growth stages of cereals. Weed. Res. 14, 415–421. doi: 10.1111/j.1365-3180.1974.tb01084.x
Zhai, Z., Xiong, K., Wang, L., Du, Y., Zhu, Z., Mao, R. (2022). Crop row detection and tracking based on binocular vision and adaptive Kalman filter. Trans. Chin. Soc Agric. Eng. 38, 143–151. doi: 10.11975/j.issn.1002-6819.2022.08.017
Zhai, Z., Zhu, Z., Du, Y., Song, Z., Mao, E. (2016a). Multi-crop-row detection algorithm based on binocular vision. Biosyst. Eng. 150, 89–103. doi: 10.1016/j.biosystemseng.2016.07.009
Zhai, Z., Zhu, Z., Du, Y., Zhang, S., Mao, E. (2016b). Method for detecting crop rows based on binocular vision with Census transformation. Trans. Chin. Soc Agric. Eng. 32, 205–213. doi: 10.11975/j.issn.1002-6819.2016.11.029
Zhang, Q., Chen, S., Li, B. (2015). Extraction method for centerlines of rice seedlings based on SUSAN corner. Trans. Chin. Soc Agric. Eng. 31, 165–171. doi: 10.11975/j.issn.1002-6819.2015.20.023
Zhang, Q., Chen, M. S., Li, B. (2017). A visual navigation algorithm for paddy field weeding robot based on image understanding. Comput. Electron. Agric. 143, 66–78. doi: 10.1016/j.compag.2017.09.008
Zhang, Q., Huang, X., Li, B. (2012). Detection of rice seedlings rows’ centerlines based on color model and nearest neighbor clustering algorithm. Trans. Chin. Soc Agric. Eng. 28, 163–171. doi: 10.3969/j.issn.1002-6819.2012.17.024
Zhang, Q., Wang, J., Li, B. (2020). Extraction method for centerlines of rice seedlings based on YOLOv3 target detection. Trans. Chin. Soc Agric. Mach. 51, 34–43. doi: 10.6041/j.issn.1000-1298.2020.08.004
Zhang, L., Zhang, R., Li, L., Ding, C., Zhang, D., Chen, L. (2022). Research on virtual Ackerman steering model based navigation system for tracked vehicles. Comput. Electron. Agric. 192, 106615. doi: 10.1016/j.compag.2021.106615
Zhao, J., Cao, D., Lan, Y., Pan, F., Wen, Y., Yang, D., et al. (2021). Extraction of maize field ridge centerline based on FCN with UAV remote sensing images. Trans. Chin. Soc Agric. Eng. 37, 72–80. doi: 10.11975/j.issn.1002-6819.2021.09.009
Zhao, R., Li, M., Zhang, M., Liu, G. (2009). Rapid crop-row detection based on improved hough transformation. Trans. Chin. Soc Agric. Mach. 40, 163–221. doi: 10.1360/972009-1549
Keywords: classical image processing, crop seedling, navigation line extraction, plantation row, growth stage
Citation: Li H, Lai X, Mo Y, He D and Wu T (2025) Pixel-wise navigation line extraction of cross-growth-stage seedlings in complex sugarcane fields and extension to corn and rice. Front. Plant Sci. 15:1499896. doi: 10.3389/fpls.2024.1499896
Received: 22 September 2024; Accepted: 30 December 2024;
Published: 30 January 2025.
Edited by:
Yanchao Zhang, Zhejiang Sci-Tech University, ChinaReviewed by:
Yuanhong Li, South China Agricultural University, ChinaCopyright © 2025 Li, Lai, Mo, He and Wu. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Tao Wu, d3Q1NXB1YkBzY2F1LmVkdS5jbg==
†Present address: Hongwei Li, Guangxi Key Laboratory of Manufacturing System and Advanced Manufacturing Technology, Guangxi University, Nanning, China
Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.
Research integrity at Frontiers
Learn more about the work of our research integrity team to safeguard the quality of each article we publish.