Skip to main content

ORIGINAL RESEARCH article

Front. Plant Sci., 11 August 2022
Sec. Sustainable and Intelligent Phytoprotection
This article is part of the Research Topic Big Data and Artificial Intelligence Technologies for Smart Forestry View all 16 articles

Automated extraction of Camellia oleifera crown using unmanned aerial vehicle visible images and the ResU-Net deep learning model

Yu Ji,,Yu Ji1,2,3Enping Yan,,Enping Yan1,2,3Xianming Yin,,Xianming Yin1,2,3Yabin SongYabin Song4Wei WeiWei Wei5Dengkui Mo,,
Dengkui Mo1,2,3*
  • 1Key Laboratory of Forestry Remote Sensing Based Big Data and Ecological Security for Hunan Province, Changsha, China
  • 2Key Laboratory of State Forestry and Grassland Administration on Forest Resources Management and Monitoring in Southern Area, Changsha, China
  • 3College of Forestry, Central South University of Forestry and Technology, Changsha, China
  • 4Central South Forest Inventory and Planning Institute of State Forestry Administration, Changsha, China
  • 5Forestry Research Institute of Guangxi Zhuang Autonomous Region, Guangxi, China

As one of the four most important woody oil-tree in the world, Camellia oleifera has significant economic value. Rapid and accurate acquisition of C. oleifera tree-crown information is essential for enhancing the effectiveness of C. oleifera tree management and accurately predicting fruit yield. This study is the first of its kind to explore training the ResU-Net model with UAV (unmanned aerial vehicle) images containing elevation information for automatically detecting tree crowns and estimating crown width (CW) and crown projection area (CPA) to rapidly extract tree-crown information. A Phantom 4 RTK UAV was utilized to acquire high-resolution images of the research site. Using UAV imagery, the tree crown was manually delineated. ResU-Net model’s training dataset was compiled using six distinct band combinations of UAV imagery containing elevation information [RGB (red, green, and blue), RGB-CHM (canopy height model), RGB-DSM (digital surface model), EXG (excess green index), EXG-CHM, and EXG-DSM]. As a test set, images with UAV-based CW and CPA reference values were used to assess model performance. With the RGB-CHM combination, ResU-Net achieved superior performance. Individual tree-crown detection was remarkably accurate (Precision = 88.73%, Recall = 80.43%, and F1score = 84.68%). The estimated CW (R2 = 0.9271, RMSE = 0.1282 m, rRMSE = 6.47%) and CPA (R2 = 0.9498, RMSE = 0.2675 m2, rRMSE = 9.39%) values were highly correlated with the UAV-based reference values. The results demonstrate that the input image containing a CHM achieves more accurate crown delineation than an image containing a DSM. The accuracy and efficacy of ResU-Net in extracting C. oleifera tree-crown information have great potential for application in non-wood forests precision management.

Introduction

Camellia oleifera, along with oil palm, coconut, and oil olive, is one of the world’s four most important woody oil-tree, and it is China’s top woody oil-tree. Camellia oleifera is widely used in cosmetics, medicine, tannin production, bio-feed, sterilization, and other fields, in addition to its primary use in the production of edible camellia oil (Zhang et al., 2013). However, the management of C. oleifera continues to rely excessively on manual labor and possesses insufficient scientific and technological support. For mastering the growth distribution of C. oleifera, rapid yield measurement, and achieving accurate management of C. oleifera forests, accurate and efficient acquisition of C. oleifera crown information is crucial (Yan et al., 2021; Ji et al., 2022).

In recent years, UAV remote sensing technology with a visible digital camera has become one of the most important methods for obtaining crop growth data due to its high resolution, real-time, and adaptability, allowing for the efficient acquisition of high-precision tree-crown data (Wang et al., 2004; Dash et al., 2019; Pearse et al., 2020; Shu et al., 2021). For crown information extraction, object-oriented classification (Zhang et al., 2015; Wu et al., 2021), watershed (Imangholiloo et al., 2019; Wu et al., 2021), local maximum (Lamar et al., 2005), and region-growing method (Pouliot and King, 2005; Bunting and Lucas, 2006) are often used. These techniques have yielded successful crown detection results for pure forests, plantations, or specific tree species and images. However, image processing parameter settings are too dependent on expert knowledge (Chadwick et al., 2020), making it difficult to automatically extract image information (Laurin et al., 2019). Therefore, new methods are required to rapidly extract tree-crown data to improve tree growth monitoring.

In recent years, image segmentation techniques based on deep learning technology that can automate and batch-process data have been widely adopted (Li et al., 2016; Kattenborn et al., 2021). Among them, the U-Net network based on Fully Convolutional Network (FCN) focuses more on segmentation details due to its capability of feature stitching and multi-scale fusion, which performs well in image segmentation (Ronneberger et al., 2015). In forestry, the U-net model has been applied successfully to tasks such as extracting tree canopy information from UAV imagery. Li et al. (2019) extracted the crown of the poplar with an accuracy of 94.1% using the U-Net network. However, if the number of U-Net network layers is excessive, network degradation will occur, and segmentation accuracy will decrease (Yang et al., 2020). The unique residual structure of the residual network can effectively mitigate the network degradation problem caused by the deep network structure and speed up network convergence (He et al., 2016). ResU-Net, which is created by combining Res-Net and U-Net, can include more layers and prevent model performance degradation (Ghorbanzadeh et al., 2021). Tong and Xu (2021) fused ResNet-34 and U-Net convolutional neural networks to create the ResNet-UNet (ResU-Net) stumpage segmentation model, improving accuracy and robustness significantly. However, the research mentioned above focuses primarily on macrophanerophytes, and there are fewer studies on the extraction of C. oleifera crown parameters. In addition, when detecting tree crowns, images with only three bands (red, green, and blue) are typically used as model input images (Neupane et al., 2019; Weinstein et al., 2019). Few studies have used multi-band images with elevation data (digital surface and canopy height models) as input images to train models for detecting tree-crown and estimating tree-crown width and projection area.

According to this context, this study is the first to apply ResU-Net to C. oleifera tree-crown extraction. UAV imagery with added elevation information (DSM and CHM) is used to create ResU-Net training datasets. This study investigates the capability of the semantic segmentation model ResU-Net to extract C. oleifera crowns from multi-band combined images with elevation information derived from UAV imagery. This study aims to (1) propose combined images with elevation information for a ResU-Net model to detect the tree-crown information of C. oleifera and (2) evaluate the models trained using various image combinations and select the optimal model for practical applications. This study is expected to provide more precise data support for the extraction of C. oleifera tree-crown information to better monitor and manage non-wood forests.

Materials and methods

Figure 1 illustrates the framework of this study. As input images, six distinct band-combined images were created first. The input and tree-crown binary images are split and amplified to obtain the training data set. The training dataset is utilized for the training of the proposed model. Then, six distinct ResU-Net models were utilized to estimate the number of plants, crown width, and projection areas at the study site. The model’s performance was then evaluated, including the precision of individual tree-crown detection, crown width, and crown projection area estimation.

FIGURE 1
www.frontiersin.org

Figure 1. Flowchart of individual tree-crown detection, crown width, and projection area assessment in this study. DSM, digital surface model; DEM, digital elevation model; CHM, canopy height model; EXG, excess green index.

Study site

The study site is located in Chenjiafang Town, Xinshao County, Shaoyang City, Hu-nan Province, between 111°08′–111°05′E and 27°15–27°38′N (Figure 2). The region has a humid mid-subtropical continental monsoon climate and average annual precipitation of 1365.2 mm, making it a typical low-hilly terrain in the south. Camellia oleifera was planted in the study area on a total area of 59.18 hm2 in 2014.

FIGURE 2
www.frontiersin.org

Figure 2. Location of the study site, Chengiafang County, Hunan, China; (A) digital surface model (DSM); (B) orthomosaic with RGB bands; (C) example of manually delineated Camellia oleifera crowns (in yellow) based on orthomosaic.

Data collection and preprocessing

Image acquisition

The UAV imagery was collected using Phantom 4 RTK1 on July 4, 2021, in diffuse light weather to avoid the influence the tree shadows have on the aerial survey results. The UAV is equipped with a 1-inch COMS sensor. The focal length of the sensor is 24 mm, the aperture range is f/2.8–f/11, and the image resolution of the camera is 5,472 × 3,648 pixels (JPEG format). The flight altitude was set to 40 m, the course speed was 3 m/s, the bypass overlap rate was set to 70%, the heading overlap rate was set to 80%, and a total of 1,127 images were captured. The UAV imagery was pre-processed using Agisoft Metashape 1.7.1 software from Agisoft LLC, Russia, which generated the digital surface model (DSM) and the RGB-banded orthomosaic. The DSM is minimally filtered (window size is 20 × 20) and then smoothed with a mean filter (window size is 5 × 5) to generate the digital elevation model (DEM), which is then subtracted from the DSM to generate the canopy height model (CHM).

The objects in the UAV imagery are primarily plants (green in color) and backgrounds (soil, rocks, plant debris, etc., which are primarily earthy in color), so the red, green, and blue bands of orthomosaic are calculated to generate EXG images (Equation 1), which are used to highlight green plants and suppress backgrounds such as shadows, rocks, and soil (Woebbecke et al., 1995).

EXG=2GRB    (1)

Where R, G, and B are the three standard bands of red, green, and blue, respectively.

Field survey data

Using UAV imagery, select 235 C. oleifera trees randomly and determine their exact location. Utilize a measuring rod to determine the height of the trees in the study area. The method is feasible because the height of C. oleifera is limited (<3 m), the tops of the trees are visible, and the distance between trees is known.

Tree-crown delineation

The tree crowns of C. oleifera were manually outlined in ArcMap 10.7 (ESRI, United States) using orthomosaic and CHM data (Table 1). There were a total of 1,862 crowns outlined (Figure 2C). Then, the manually delineated tree-crown image is binarized, the background pixel value is changed to 0, the pixel value of the C. oleifera crown area is changed to 255, and a tree-crown binary image is generated that corresponds to the tree-crown in the UAV image.

TABLE 1
www.frontiersin.org

Table 1. Statistics of crown width and crown projection area.

The CV2 function provided by OpenCV, an open-source computer vision library, was used to count the number of pixels contained in the tree-crown of each of the 235 trees (Section “Field survey data”) based on the tree-crown mask image and calculate the crown projection area (CPA) based on the image resolution. Using Canny’s algorithm (Hu et al., 2018) to extract the edge features of each crown of 235 trees, followed by the ellipse fitting algorithm (Yan et al., 2008) to obtain each crown’s external ellipse. Calculate the long and short axes of the ellipse as the maximum and minimum values of C. oleifera crown width, and then calculate the mean of these two values to obtain the average crown width of C. oleifera (Zhang et al., 2021). The crown width and projection area of 235 trees were calculated based on the 0.01532703 m image resolution.

Dataset preparation

Input image

The blue, green, red, and EXG products are used as input images for the division of the tree crown. The CHM or DSM was added to the combined images to compare the effects of input images containing different elevation information on the model’s ability to accurately estimate the crown width and projection area (Table 2).

TABLE 2
www.frontiersin.org

Table 2. Combination of different images for ResU-Net training.

Training dataset

Thousand eight hundred and sixty-two delineated trees (Section “Tree crown delineation”) were separated into a training and validation set and a test set for this study. First, the six-band combination images containing information about the tree crowns of 1,627 trees and the corresponding binary tree-crown images are divided into 256 × 256 pixel image tiles for processing. In addition, the training data are rotated by 90°, 180°, and 270° from its original orientation to increase the number of training samples and enhance the model’s robustness. In summary, six training datasets, each one containing 3,375 images, were obtained for this study.

To evaluate the performance of the final model, 235 C. oleifera trees with crown width and projection area reference information from UAV imagery were used as the test set.

ResU-Net model

Since C. oleifera tree-crown images contain a large number of background interference factors, such as weeds and soil, ResNet101, which has a strong feature extraction capability (Laurin et al., 2019), is used as the backbone network and combined with the U-Net (Ronneberger et al., 2015) network design concept to create the ResU-Net network model in this study. ResU-Net presents the Residual Block (Res-block) structure (illustrated in Figure 3B) based on the U-Net network, which can effectively overcome the network degradation and gradient dispersion issues caused by an increase in network depth and accelerate network convergence (Wu et al., 2019). ResU-Net requires a smaller training set and focuses more on image segmentation details without compromising accuracy (Yang et al., 2019), and it can recognize the crown of C. oleifera at the pixel level.

FIGURE 3
www.frontiersin.org

Figure 3. ResU-Net network structure: (A) ResU-Net network framework; (B) Res-block modules.

The encoder (to the left of the dashed line) and decoder (right of the dashed line) make up the ResU-Net structure, as shown in Figure 3A. The encoder is used to downsample the input image, capture image context information, and extract image semantic information features. The decoder upsamples the image using transposed convolution and concatenates features of the same dimension to provide detailed feature information (Chen et al., 2021).

For the encoder, the input image passes through a Convolution layer (CONV) with a convolution kernel of 7 × 7 and a step size of 2, followed by a Max Pooling Layer. The image size is decreased to a quarter of its original size, and the number of channels is increased to 64. Following this, the tree-crown features are extracted through the residual block (Res-block) until an 8 × 8 feature vector with a depth of 2048 is obtained.

The initial step for the decoder is to increase the size of the feature layer and decrease its depth by upsampling. Next, the downsampled and upsampled feature layers of equal size are concatenated. The concatenated feature layers are fused using a 3 × 3 convolutional layer, a Batch Normalization layer (BN), and a Rectified linear unit after each concatenation (ReLU). The tree-crown binary mask of C. oleifera is obtained through a final convolution operation.

The hold-out method is selected for cross-validation, and the mIoU of the model is set to stop in advance without increasing within 10 epochs to prevent overfitting of the model. Only the model with improved accuracy after each training is saved. Res-Net weights pre-trained on ImageNet are used for transfer learning. The ResU-Net model was then trained using the six training datasets. Set the learning rate parameter to 0.001, the epochs to 100, and the batch size to 4 when training the model. Based on these six distinct composite images, six ResU-Net models were generated. All models were trained on a Windows 10 desktop with an Intel i7 6700k CPU, 6 GB GPU, and 24 GB RAM using PyCharm 2010.1.4 software based on the Pytorch framework for deep learning.

Accuracy evaluation

To determine the optimal ResU-Net model, the accuracy of tree detection, crown width, and projection area assessment were calculated separately. The test set, which contained crown width and projection area reference data from the UAV images, was utilized to assess the performance of each model.

The intersection over union (IoU) was used to assess the ResU-Net model’s accuracy in detecting tree crowns (Equation 2). IoU measures the area of the union and intersection of the crown polygons of the test set and the crown polygons predicted by the model. When IoU exceeded 50%, it was deemed acceptable. ResU-Net model’s detection of individual tree crowns was evaluated using precision, recall, and F1 score (Equations 35). Precision is the proportion of accurately detected trees in a model detection. The recall is the proportion of correctly identified trees in the test set. The F1 score indicates the overall test accuracy, which is based on recall and precision (Shao et al., 2019; Hao et al., 2021).

IoU=AactualApredictedAactualApredicted×100%    (2)
P=TPTP+FP×100%    (3)
R=TPTP+FN×100%    (4)
F1=2PRP+R×100%    (5)

where TP is the number of correctly identified trees by the model and IoU is >50%. FN is the number of trees omitted by the model when the IoU is <50%, and FP is the number of other tree or weed species found. The term Aactual refers to the crown polygons of the test set. The crown polygons predicted by the ResU-Net model are indicated by Apredicted. The intersection operation represents the area shared by Aactual and Apredicted, whereas the union operation represents the area formed when Aactual and Apredicted are combined.

The 235 tree-crown widths and projection area from the UAV imagery were then compared to the six ResU-Net model predictions. Coefficient of determination (R2), root mean square error (RMSE), and relative RMSE (rRMSE) were utilized to evaluate the model’s accuracy in estimating tree-crown width and projection area (Equations 68). R2 is utilized to assess fitness, while RMSE and rRMSE are employed to estimate error.

Finally, a comprehensive analysis of the accuracy of individual tree-crown detection, crown width, and predicted area assessment was conducted to determine the best ResU-Net model for crown width and crown projection area assessment.

R2=1i=1N(yixi)2i=1N(yiy¯i)2    (6)
RMSE=i=1N(yixi)2N    (7)
rRMSE=RMSExi¯    (8)

where N represents the number of trees that the model has detected. yi denotes the crown width and predicted area from the assessed datasets. y¯i represents the average crown width and predicted area from the assessed datasets; xi denotes the crown width and predicted area from the reference dataset; and x¯i represents the average crown width and predicted area from the reference dataset.

Results

Detection of individual tree crown

Figure 4 illustrates examples of the ResU-Net model identifying individual tree crowns at the study site. The accuracy of crown detection and delineation is shown in Table 3. RGB-CHM had the highest precision (88.73%) for tree detection, followed by RGB-DSM (87.92%), RGB (85.00%), EXG-CHM (82.30%), EXG-DSM (80.71%), and EXG had the lowest precision (80.68%). The RGB-CHM combination had the highest recall rate (80.43%), while the EXG-DSM combination had the lowest (67.66%). The F1 score of the RGB-CHM combination reached 84.38%, followed by RGB-DSM (82.35%), RGB (82.20%), EXG-CHM (77.48%), EXG (75.57%), and EXG-DSM (73.61%). The RGB-CHM combination had the highest IoU (91.38%) for tree-crown delineation, while the EXG-DSM combination had the lowest (87.07%).

FIGURE 4
www.frontiersin.org

Figure 4. Example of tree-crown detection and segmentation. (A) original image; (B) manually delineated result; (C–H) are the crown detection results of the ResU-Net model based on the combination of RGB, RGB-CHM, RGB-DSM, EXG, EXG-CHM, and EXG-DSM.

TABLE 3
www.frontiersin.org

Table 3. The accuracy of individual tree-crown detection using different combinations.

The results of all ResU-Net models yielded precision >80%, recall >65%, F1 score >70%, and IoU >85%. When the input image is based on the RGB or EXG combinations of ResU-Net and contains CHM, the precision, recall, F1 score, and IoU are greater than when the input image contains DSM, reflecting the higher precision of tree-crown detection and delineation in the CHM combination. The model’s average processing time for tree-crown detection in each image is 0.16 s, which meets the requirements of practical applications.

Extraction of tree-crown width

Using the UAV-based tree-crown width as reference. As a result of their proximity to the study area’s boundaries, certain canopies were omitted from the evaluation of the model’s accuracy in estimating tree-crown width, resulting in incomplete shapes. The accuracy of the ResU-Net model in estimating the tree-crown width based on RGB, RGB-CHM, RGB-DSM, EXG, EXG-CHM, and EXG-DSM as input images, respectively, is illustrated in Figure 5.

FIGURE 5
www.frontiersin.org

Figure 5. Linear regressions of tree-crown width between UAV imagery and different ResU-Net models. (A) RGB; (B) RGB-CHM; (C) RGB-DSM; (D) EXG; (E) EXG-CHM; (F) EXG-DSM. The dotted line represents a 1:1 match, and the red line represents the trend of the tree-crown width relationship based on the ResU-Net model and UAV imagery.

All the results of the six models for estimating tree-crown width yielded R2 > 0.70. The tree-crown width estimation based on the RGB combinations yielded higher accuracy (R2 > 0.89, 6.47% ≤ rRMSE ≤ 7.48%). The accuracy based on the EXG combinations is lower (R2 < 0.84, 9.15% ≤ rRMSE ≤ 11.51%). Among them, the accuracy of estimating tree-crown width using the DSM combination was the lowest, while the CHM combination produced better results, and the RGB-CHM combination achieved the highest accuracy (R2 = 0.9271, RMSE = 0.1282 m, rRMSE = 6.47%).

Extraction of tree-crown projection area

The projected tree-crown area estimated by six distinct ResU-Net models was compared to the UAV-based crown projection area (Figure 6). The accuracy of tree-crown projection area estimation based on the RGB combinations (R2 > 0.92, 9.39% ≤ rRMSE ≤ 11.70%) was higher than that based on the EXG combinations (R2 < 0.85, 15.04% ≤ rRMSE ≤ 19.06%). Among them, the CHM model was more accurate than the DSM model. The model with the RGB-CHM combination produced the highest accuracy (R2 = 0.9498, RMSE = 0.2675 m2, rRMSE = 9.39%), while the model with the EXG combination produced the lowest accuracy, which is comparable to the tree-crown width from the UAV imagery. ResU-Net model predictions were lower compared to the UAV-based crown width and crown projection area because of the overlapping and shading conditions with unclear boundaries.

FIGURE 6
www.frontiersin.org

Figure 6. Linear regressions of crown projection area between UAV imagery and different ResU-Net models. (A) RGB; (B) RGB-CHM; (C) RGB-DSM; (D) EXG; (E) EXG-CHM; (F) EXG-DSM. The dotted line represents a 1:1 match, and the red line represents the trend of the crown projection area relationship based on the ResU-Net model and UAV imagery.

The ResU-Net model accurately estimated the tree-crown width and crown projection area. RGB was more accurate than EXG when it came to modeling accuracy. The accuracy is greater when the input image of ResU-Net-based RGB or EXG combinations contains CHM than when the input image contains DSM. Comparing the accuracy of individual tree-crown detection, tree-crown width, and projection area between models, the RGB-CHM combination was the optimal combination for the ResU-Net model’s detection of tree-crown width and projection area.

Discussion

This study proposes to use the combined images and elevation data from UAV imagery to create datasets for training ResU-Net models to automatically extract C. oleifera tree-crown and estimate crown width and crown projection area parameters. The results demonstrate that the trained ResU-Net model can detect tree crowns and accurately estimate crown width and crown projection area. The ResU-Net model has excellent generalizability and high stability, allowing it to fulfill the need for C. oleifera tree-crown data in agricultural production.

Performance of the model

Individual tree-crown detection

The optimal ResU-Net model based on the RGB-CHM combination achieved high accuracy for tree-crown detection (precision = 88.73%, recall = 80.43%, F1 score = 84.68%, and IoU = 91.38%). Hao et al. (2021) and Braga et al. (2020) used the Mask-RCNN model to detect macrophanerophyte canopies, yielding F1scores of 84.68% and 86%, which are comparable to the F1-score of this study, whereas the IoU values yielded 91.27% and 61% are smaller than the IoU of this study, respectively. Jin et al. (2020) reported that the F1-score of 74.04% and accuracy of 79.45% for tree-crown detection based on U-Net and marker-control watershed algorithm, which is lower than the F1 score and accuracy of this study.

Next, this study compares the performance of the ResU-Net with the classical watershed algorithm, the U-Net model, the U-Net++ model (Zhou et al., 2018), and the DeepLabV3 Plus model (Chen et al., 2018) for tree-crown detection (Figure 7). According to the crown detection results, the crown detection accuracy based on the deep learning method is significantly higher than that of the classical watershed algorithm. Since C. oleifera trees are lower and there is interference from grass and other tree species, etc., the watershed algorithm is less precise. The accuracy of tree-crown detection using the ResU-Net model was slightly higher than that of the U-Net (precision = 86.18%, recall = 79.57%, F1 score = 82.74%, and IoU = 90.75%) and U-Net++ model (precision = 87.50%, recall = 77.45%, F1 score = 82.17%, and IoU = 90.96%). Using the DeepLabV3 Plus model for tree-crown detection yielded the accuracy (precision = 89.41%, recall = 79.72%, F1 score = 84.29%, and IoU = 91.85%) comparable to the ResU-Net model. In conclusion, the detection accuracy of different network models for tree crowns is similar, indicating that the application of deep learning methods for extracting C. oleifera tree crowns from UAV visible images is universal, and the accuracy is generally high. With the advancement of deep learning techniques, we can utilize more robust network models for C. oleifera tree-crown detection in the future study.

FIGURE 7
www.frontiersin.org

Figure 7. Comparison of tree-crown detection results between ResU-Net and other methods. (A) original image; (B) manually delineated result; (C–G) are examples of tree-crown detection using the ResU-Net model, DeepLabV3+ model, watershed algorithm, U-Net model, and U-Net++ model.

Crown width assessment

The key innovation of this study is to use the ResU-Net model and combine images with elevation data to estimate crown width. The RGB-CHM combination provided the most accurate measurements of crown width (R2 = 0.9271, RMSE = 0.1282 m, rRMSE = 6.47%). Few studies combine the ResU-Net model with elevation data to estimate the crown width of C. oleifera. Consequently, the accuracy of crown width and projection area estimation in our study is compared to that of studies employing alternative remote sensing techniques. Wu et al. (2021) used the optimized watershed with multi-scale markers method to estimate C. oleifera crown width yielded R2 = 0.75. Dong et al. (2020) estimated the crown width of apple and pear trees using local maximum and marker-controlled watershed algorithms, yielding R2 values of 0.78 and 0.68, respectively. Compared to these conventional remote sensing techniques (e.g., watershed, local maximum algorithms), the present method has greater accuracy in estimating crown width, and it can be automated.

Crown projection area assessment

The RGB-CHM combination provided the most accurate measurements of crown projection area (R2 = 0.9498, RMSE = 0.2675 m2, and rRMSE = 9.39%). Mu et al. (2018) estimated the peach crown projection area using adaptive thresholding and marker-controlled watershed segmentation with R2 = 0.89 and RMSE = 3.87 m2. Dong et al. (2020) estimated the crown projection area of apple and pear trees using local maximum and marker-controlled watershed algorithms, yielding R2 values of 0.87 and 0.81, respectively. Ye et al. (2022) estimated the olive crown projection area using the U2-Net model, producing R2 > 0.93, which is comparable to our study. However, it yielded MRE = 14.27% higher than our study (MRE = 12.23%).

Next, compared to the crown projection area extracted using RGB (R2 = 0.9220), R2 increased by 0.0278 with the addition of CHM, and by 0.0059 with the addition of DSM, respectively. Compared to the crown width extracted using RGB (R2 = 0.8936), R2 increased by 0.0335 with the addition of CHM, and by 0.0146 with the addition of DSM, respectively. As can be seen, the addition of CHM to the model has a greater impact on the prediction accuracy of the CW and CPA than the addition of DSM. However, the increased value (<0.05) of the model accuracy after adding the elevation information is low, because the C. oleifera planting areas are mainly hilly with little elevation change, and the C. oleifera are shrubs with low tree height. To verify the reliability of the experimental results of this study, it is necessary to conduct in-depth experiments in a region with a large height difference in the future study.

In addition, this study discovered that the accuracy of the EXG combinations was lower than that of the RGB combinations, indicating that the limited features of EXG (grayscale maps) reduce the model’s accuracy (Hao et al., 2021). Therefore, it is recommended that adequate features are included in the training process.

Factors of influence on model performance

Accuracy of CHM extraction

The accuracy of canopy height model (CHM) extraction influences the model-detected tree crown, the estimated crown width, and the crown projection area. To evaluate the accuracy of CHM, the UAV-derived tree height estimates were compared with field measurements in this study.

In this study, the CHM and the manually outlined tree-crown vector map were combined. Then, ArcMap 10.7 was utilized to extract the maximum value of CHM for each tree-crown region and estimate the tree height from the UAV imagery. For the accuracy of the results, incomplete C. oleifera crowns at the image margins were omitted. There were 183 complete canopies that corresponded to the estimated tree height values when measured on the ground. As depicted in Figure 8, the correlation between the estimated tree height and field measurements was strong, with R2 of 0.7853, RMSE of 0.2688 m, and rRMSE of 17.64%. The estimated tree heights were lower than field measurements, likely due to the high density of C. oleifera forests and lack of bare ground, which prevented the filtering method from obtaining sufficient ground area when DSM was processed. Consequently, the digital elevation model (DEM) is higher than the true elevation value, whereas the CHM is lower than the true CHM.

FIGURE 8
www.frontiersin.org

Figure 8. Linear regressions of tree height between the field measurement and different ResU-Net models.

Airborne laser scanning (ALS) has a high penetrability and is considered the best option for tree detection with high accuracy in dense canopy areas (Sankey et al., 2017; Pourshamsi et al., 2021). Therefore, orthophotos and ALS can be combined as input data when training the model.

Spatial image resolution

The spatial resolution of the image has a significant impact on the accuracy of the model’s detection of tree crowns. By comparing 0.3, 1.5, 2.7, and 6.3 cm spatial resolutions, Fromm et al. (2019) determined that the model had the highest average accuracy for detecting conifer seedlings when the spatial resolution was 0.3 cm. Schiefer et al. (2020) concluded that, when using five spatial resolution images for crown detection, the lower the spatial resolution, the lower the model detection accuracy. Studies have demonstrated that the higher the spatial resolution, the more precise the model’s crown detection. However, when the resolution exceeds a certain threshold, it does not improve the accuracy of models significantly (Hao et al., 2021). The model’s accuracy is also dependent on the size of the tree crowns, and the ratio of the tree-crown diameter to the spatial resolution is a crucial factor in determining the detection accuracy. Yin and Wang (2019) suggested that images with a spatial resolution greater than a quarter of the crown diameter had the highest accuracy for crown detection; however, 0.25 m resolution had the highest accuracy for mangrove crowns compared to 0.1, 0.5, and 1 m resolutions. An excessively high spatial resolution will generate interference due to excessive detail and noise, which is not conducive to model crown detection.

In addition, the measurement error resulting from manual delineation and the calculation error of crown parameters will also have an impact on the model estimation results. For optimal tree-crown extraction results, the measurement method must be optimized. The ratio relationship between spatial image resolution and tree-crown size must be further investigated in a subsequent study.

Conclusion

Combining the ResU-Net model with images that add elevation information (CHM or DSM) from UAV imagery can effectively and automatically detect C. oleifera tree crowns and estimate the crown width and crown projection area, which has significant application potential. ResU-Net model with RGB-CHM combination outperformed other models with different combinations (Precision = 88.73%, IoU = 91.38%, Recall = 80.43%, and F1 score = 84.38%). The model’s accuracy using RGB combinations was superior to the model’s accuracy using EXG combinations. The accuracy of crown width and crown projection area estimation is dependent on the input elevation data (DSM or CHM), and the model with CHM data is more accurate. In this study, the ResU-Net model with RGB-CHM combination provided the most accurate estimates of crown width (R2 = 0.9271, RMSE = 0.1282 m, rRMSE = 6.47%) and crown projection area (R2 = 0.9498, RMSE = 0.2675 m2, rRMSE = 9.39%). In conclusion, the combination of the deep learning model ResU-Net and UAV images containing elevation information has great potential for extracting crown information from C. oleifera. This method can obtain high-precision information on the tree crowns of C. oleifera trees at a low cost and with a high degree of efficiency, making it ideal for the precise management and rapid yield estimation of C. oleifera forests.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Author contributions

YJ and DM: conceptualization. YJ: data curation and writing—original draft. YJ and XY: investigation and methodology. DM, YS and WW: resources and supervision. XY: validation. YJ and EY: writing—reviewing and editing. All authors contributed to the article and approved the submitted version.

Funding

This research was funded by the National Natural Science Foundation of China (grant numbers: 31901311 and 32071682) and Science and Technology Innovation Plan Project of Hunan Provincial Forestry Department (grant number XLK202108-8).

Acknowledgments

The authors would like to thank the Key Laboratory of Forestry Remote Sensing Big Data and Ecological Security for Hunan Province for providing the Phantom 4 RTK UAV. We would also like to thank the staff of the Camellia oleifera plantation. We would like to thank MogoEdit (https://www.mogoedit.com) for its English editing during the preparation of this manuscript.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Footnotes

References

Braga, J. R., Peripato, V., Dalagnol, R., Ferreira, M. P., Tarabalka, Y., Aragão, L. E., et al. (2020). Tree crown delineation algorithm based on a convolutional neural network. Remote Sens. 12:1288. doi: 10.3390/rs12081288

CrossRef Full Text | Google Scholar

Bunting, P., and Lucas, R. (2006). The delineation of tree crowns in Australian mixed species forests using hyperspectral compact airborne spectrographic imager (CASI) data. Remote Sens. Environ. 101, 230–248. doi: 10.1016/j.rse.2005.12.015

CrossRef Full Text | Google Scholar

Chadwick, A. J., Goodbody, T. R. H., Coops, N. C., Hervieux, A., Bater, C. W., Martens, L. A., et al. (2020). Automatic delineation and height measurement of regenerating conifer crowns under leaf-Off conditions using UAV imagery. Remote Sens. 12:4104. doi: 10.3390/rs12244104

CrossRef Full Text | Google Scholar

Chen, Y., Zhu, C., Hu, X., and Wang, L. (2021). Detection of wheat stem section parameters based on improved Unet. Trans. Chin. Soc. Agric. Mach. 52, 169–176. doi: 10.6041/j.issn.1000-1298.2021.07.017

CrossRef Full Text | Google Scholar

Chen, L. C., Zhu, Y., Papandreou, G., Schroff, F., and Adam, H. (2018). “Encoder-decoder with atrous separable convolution for semantic image segmentation.” in Proceedings of the European conference on computer vision (ECCV). September 8–14, 2018; 801–818.

Google Scholar

Dash, J. P., Watt, M. S., Paul, T. S. H., Morgenroth, J., and Pearse, G. D. (2019). Early detection of invasive exotic trees using UAV and manned aircraft multispectral and LiDAR data. Remote Sens. 11:1812. doi: 10.3390/rs11151812

CrossRef Full Text | Google Scholar

Dong, X., Zhang, Z., Yu, R., Tian, Q., and Zhu, X. (2020). Extraction of information about individual trees from high-spatial-resolution UAV-acquired images of an orchard. Remote Sens. 12:133. doi: 10.3390/rs12010133

CrossRef Full Text | Google Scholar

Fromm, M., Schubert, M., Castilla, G., Linke, J., and McDermid, G. (2019). Automated detection of conifer seedlings in drone imagery using convolutional neural networks. Remote Sens. 11:2585. doi: 10.3390/rs11212585

CrossRef Full Text | Google Scholar

Ghorbanzadeh, O., Crivellari, A., Ghamisi, P., Shahabi, H., and Blaschke, T. (2021). A comprehensive transferability evaluation of U-net and ResU-net for landslide detection from Sentinel-2 data (case study areas from Taiwan, China, and Japan). Sci. Rep. 11, 14629–14620. doi: 10.1038/s41598-021-94190-9

PubMed Abstract | CrossRef Full Text | Google Scholar

Hao, Z., Lin, L., Post, C. J., Mikhailova, E. A., Li, M., Yu, K., et al. (2021). Automated tree-crown and height detection in a young forest plantation using mask region-based convolutional neural network (mask R-CNN). ISPRS-J. Photogramm. Remote Sens. 178, 112–123. doi: 10.1016/j.isprsjprs.2021.06.003

CrossRef Full Text | Google Scholar

He, K., Zhang, X., Ren, S., and Sun, J. (2016). “Deep residual learning for image recognition.” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, June 27–30, 2016, 770–778.

Google Scholar

Hu, T., Qi, Z., Qin, Y., and Shi, J. (2018). “Detecting technology of current generator blade shape based on improved canny algorithm.” in IOP Conference Series: Materials Science and Engineering; September 15–16, 2018. Vol. 452 (IOP Publishing), 32026.

Google Scholar

Imangholiloo, M., Saarinen, N., Markelin, L., Rosnell, T., Näsi, R., Hakala, T., et al. (2019). Characterizing seedling stands using leaf-Off and leaf-On photogrammetric point clouds and Hyperspectral imagery acquired from unmanned aerial vehicle. Forests 10:415. doi: 10.3390/f10050415

CrossRef Full Text | Google Scholar

Ji, Y., Yin, X., Yan, E., Jiang, J., Peng, S., and Mo, D. (2022). Research on extraction of shape features of camellia oleifera fruit based on camera photography. J. Nanjing For. Univ. (Nat Sci Ed) 46, 63–70. doi: 10.12302/j.issn.1000-2006.202109030

CrossRef Full Text | Google Scholar

Jin, Z., Cao, S., Wang, L., and Sun, W. (2020). A method for individual tree-crown extraction from UAV remote sensing image based on U-net and watershed algorithm. J. Northwest For. Univ. 35, 194–204. doi: 10.3969/j.issn.1001-7461.2020.06.27

CrossRef Full Text | Google Scholar

Kattenborn, T., Leitloff, J., Schiefer, F., and Hinz, S. (2021). Review on convolutional neural networks (CNN) in vegetation remote sensing. ISPRS J. Photogramm. Remote Sens. 173, 24–49. doi: 10.1016/j.isprsjprs.2020.12.010

CrossRef Full Text | Google Scholar

Lamar, W. R., McGraw, J. B., and Warner, T. A. (2005). Multitemporal censusing of a population of eastern hemlock (Tsuga canadensis L.) from remotely sensed imagery using an automated segmentation and reconciliation procedure. Remote Sens. Environ. 94, 133–143. doi: 10.1016/j.rse.2004.09.003

CrossRef Full Text | Google Scholar

Laurin, G. V., Ding, J., Disney, M., Bartholomeus, H., Herold, M., Papale, D., et al. (2019). Tree height in tropical forest as measured by different ground, proximal, and remote sensing instruments, and impacts on above ground biomass estimates. Int. J. Appl. Earth Obs. Geoinf. 82:101899. doi: 10.1016/j.jag.2019.101899

CrossRef Full Text | Google Scholar

Li, W., Fu, H., Yu, L., and Cracknell, A. (2016). Deep learning based oil palm tree detection and counting for high-resolution remote sensing images. Remote Sens. 9:22. doi: 10.3390/rs9010022

CrossRef Full Text | Google Scholar

Li, Y. S., Zhen, H. W., and Luo, G. P. (2019). Extraction and counting of pop-ulus Euphratica crown using UAV images integrated with U-Netmethod. Remote Sens. Technol. Appl. 34, 939–949. doi: 10.11873/j.issn.1004-0323.2019.5.0939

CrossRef Full Text | Google Scholar

Mu, Y., Fujii, Y., Takata, D., Zheng, B., Noshita, K., Honda, K., et al. (2018). Characterization of peach tree crown by using high-resolution images from an unmanned aerial vehicle. Hortic. Res. 5:74. doi: 10.1038/s41438-018-0097-z

PubMed Abstract | CrossRef Full Text | Google Scholar

Neupane, B., Horanont, T., and Hung, N. D. (2019). Deep learning based banana plant detection and counting using high-resolution red-green-blue (RGB) images collected from unmanned aerial vehicle (UAV). PLoS One 14:e0223906. doi: 10.1371/journal.pone.0223906

PubMed Abstract | CrossRef Full Text | Google Scholar

Pearse, G. D., Tan, A. Y., Watt, M. S., Franz, M. O., and Dash, J. P. (2020). Detecting and mapping tree seedlings in UAV imagery using convolutional neural networks and field-verified data. ISPRS J. Photogram. Remote Sens. 168, 156–169. doi: 10.1016/j.isprsjprs.2020.08.005

CrossRef Full Text | Google Scholar

Pouliot, D., and King, D. (2005). Approaches for optimal automated individual tree crown detection in regenerating coniferous forests. Can. J. For. Res. 31, 255–267. doi: 10.5589/m05-011

CrossRef Full Text | Google Scholar

Pourshamsi, M., Xia, J., Yokoya, N., Garcia, M., Lavalle, M., Pottier, E., et al. (2021). Tropical forest canopy height estimation from combined polarimetric SAR and LiDAR using machine-learning. ISPRS J. Photogramm. Remote Sens. 172, 79–94. doi: 10.1016/j.isprsjprs.2020.11.008

CrossRef Full Text | Google Scholar

Ronneberger, O., Fischer, P., and Brox, T. (2015). U-net: convolutional networks for biomedical image segmentation. Proc. Med. Image Comput. Comput. Assist. Intervention 9351, 234–241. doi: 10.1007/978-3-319-24574-4_28

CrossRef Full Text | Google Scholar

Sankey, T., Donager, J., McVay, J., and Sankey, J. B. (2017). UAV LiDAR and hyperspectral fusion for forest monitoring in the southwestern USA. Remote Sens. Environ. 195, 30–43. doi: 10.1016/j.rse.2017.04.007

CrossRef Full Text | Google Scholar

Schiefer, F., Kattenborn, T., Frick, A., Frey, J., Schall, P., Koch, B., et al. (2020). Mapping forest tree species in high resolution UAV-based RGB-imagery by means of convolutional neural networks. ISPRS J. Photogramm. Remote Sens. 170, 205–215. doi: 10.1016/j.isprsjprs.2020.10.015

CrossRef Full Text | Google Scholar

Shao, Z., Pan, Y., Diao, C., and Cai, J. (2019). Cloud detection in remote sensing images based on multiscale features-convolutional neural network. IEEE Trans. Geosci. Remote Sens. 57, 4062–4076. doi: 10.1109/TGRS.2018.2889677

CrossRef Full Text | Google Scholar

Shu, M., Li, S., Wei, J., Che, Y., Li, B., and Ma, Y. (2021). Extraction of citrus crown parameters using UAV platform. Trans. CSAE 37, 68–76. doi: 10.11975/j.issn.1002-6819.2021.01.009

CrossRef Full Text | Google Scholar

Tong, Z., and Xu, A. J. (2021). A tree segmentation method based on ResNet-UNet. J. Cent. South Univ. For. Technol. 41, 132–139. doi: 10.14067/j.cnki.1673-923x.2021.01.014

CrossRef Full Text | Google Scholar

Wang, L., Gong, P., and Biging, G. S. (2004). Individual tree-crown delineation and treetop detection in high-spatial-resolution aerial imagery. Photogramm. Eng. Remote. Sens. 70, 351–357. doi: 10.14358/PERS.70.3.351

CrossRef Full Text | Google Scholar

Weinstein, B. G., Marconi, S., Bohlman, S., Zare, A., and White, E. (2019). Individual tree-crown detection in RGB imagery using semi-supervised deep learning neural networks. Remote Sens. 11:1309. doi: 10.3390/rs11111309

CrossRef Full Text | Google Scholar

Woebbecke, D., Meyer, G., von Bargen, K., and Mortensen, D. (1995). Color indices for weed identification under various soil, residue and lighting conditions. Ame. Soc. Agric. Eng. 38, 259–269. doi: 10.13031/2013.27838

CrossRef Full Text | Google Scholar

Wu, J., Peng, S., Jiang, F., Tang, J., and Sun, H. (2021). Extraction of Camellia oleifera crown width based on the method of optimized watershed with multi-scale markers. Chin. J. Appl. Ecol. 32, 2449–2457. doi: 10.13287/j.1001–9332.202107.010

CrossRef Full Text | Google Scholar

Wu, Z., Shen, C., and van den Hengel, A. (2019). Wider or deeper: revisiting the resnet model for visual recognition. Pattern Recognit. 90, 119–113. doi: 10.1016/j.patcog.2019.01.006

CrossRef Full Text | Google Scholar

Yan, E., Ji, Y., Yin, X., and Mo, D. (2021). Rapid estimation of Camellia oleifera yield based on automatic detection of canopy fruits using UAV images. Trans. CSAE 37, 39–46. doi: 10.11975/j.issn.1002-6819.2021.16.006

CrossRef Full Text | Google Scholar

Yan, B., Wang, B., and Li, Y. (2008). Optimal ellipse fitting method based on least-square principle. J. Beijing Univ. Aeronaut. Astronaut. 34, 295–298. doi: 10.13700/j.bh.1001-5965.2008.03.001

CrossRef Full Text | Google Scholar

Yang, J., Faraji, M., and Basu, A. (2019). Robust segmentation of arterial walls in intravascular ultrasound images using dual path U-net. Ultrasonics 96, 24–33. doi: 10.1016/j.ultras.2019.03.014

PubMed Abstract | CrossRef Full Text | Google Scholar

Yang, T., Song, J., Li, L., and Tang, Q. (2020). Improving brain tumor segmentation on MRI based on the deep U-net and residual units. J. Xray Sci. Technol. 28, 95–110. doi: 10.3233/XST-190552

PubMed Abstract | CrossRef Full Text | Google Scholar

Ye, Z., Wei, J., Lin, Y., Guo, Q., Zhang, J., Zhang, H., et al. (2022). Extraction of olive crown based on UAV visible images and the U2-net deep learning model. Remote Sens. 14:1523. doi: 10.3390/rs14061523

CrossRef Full Text | Google Scholar

Yin, D., and Wang, L. (2019). Individual mangrove tree measurement using UAV-based LiDAR data: possibilities and challenges. Remote Sens. Environ. 223, 34–49. doi: 10.1016/j.rse.2018.12.034

CrossRef Full Text | Google Scholar

Zhang, D. S., Jin, Q. Z., Xue, Y. L., Zhang, D., and Wang, X. G. (2013). Nutritional value and adulteration identification of oil tea camellia seed oil. China Oils and Fats. 38, 47–50. doi: 10.3969/j.issn.1003-7969.2013.08.013

CrossRef Full Text | Google Scholar

Zhang, N., Feng, Y. W., Zhang, X. L., and Fan, J. C. (2015). Extracting individual tree crown by combining spectral and texture features from aerial images. J. Beijing For. Univ. 37, 13–19. doi: 10.13332/j.1000-1522.20140309

CrossRef Full Text | Google Scholar

Zhang, Y., Zhang, C., Wang, J., Li, H., Ai, M., and Yang, A. (2021). Individual tree crown width extraction and DBH estimation model based on UAV remote sensing. For. Resour. Manage. 0, 67–75. doi: 10.13466/j.cnki.lyzygl.2021.03.011

CrossRef Full Text | Google Scholar

Zhou, Z., Rahman Siddiquee, M. M., Tajbakhsh, N., and Liang, J. (2018). “Unet++: a nested u-net architecture for medical image segmentation,” in Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support. ed. S. Danail (Cham: Springer), 3–11.

Google Scholar

Keywords: UAV imagery, deep learning, image segmentation, tree-crown detection, Camellia oleifera

Citation: Ji Y, Yan E, Yin X, Song Y, Wei W and Mo D (2022) Automated extraction of Camellia oleifera crown using unmanned aerial vehicle visible images and the ResU-Net deep learning model. Front. Plant Sci. 13:958940. doi: 10.3389/fpls.2022.958940

Received: 01 June 2022; Accepted: 25 July 2022;
Published: 11 August 2022.

Edited by:

Weipeng Jing, Northeast Forestry University, China

Reviewed by:

William Zhao, Columbia University, United States
Aichen Wang, Jiangsu University, China

Copyright © 2022 Ji, Yan, Yin, Song, Wei and Mo. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Dengkui Mo, dengkuimo@csuft.edu.cn

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.