Skip to main content

ORIGINAL RESEARCH article

Front. Plant Sci., 15 August 2023
Sec. Sustainable and Intelligent Phytoprotection
This article is part of the Research Topic Pests and Diseases Monitoring and Forecasting Algorithms, Technologies, and Applications View all 19 articles

Skip DETR: end-to-end Skip connection model for small object detection in forestry pest dataset

Bing Liu,Bing Liu1,2Yixin JiaYixin Jia2Luyang LiuLuyang Liu2Yuanyuan DangYuanyuan Dang2Shinan Song*Shinan Song2*
  • 1College of Computer Science and Technology, Jilin University, Changchun, Jilin, China
  • 2School of Computer Science and Engineering, Changchun University of Technology, Changchun, Jilin, China

Object detection has a wide range of applications in forestry pest control. However, forest pest detection faces the challenges of a lack of datasets and low accuracy of small target detection. DETR is an end-to-end object detection model based on the transformer, which has the advantages of simple structure and easy migration. However, the object query initialization of DETR is random, and random initialization will cause the model convergence to be slow and unstable. At the same time, the correlation between different network layers is not strong, resulting in DETR is not very ideal in small object training, optimization, and performance. In order to alleviate these problems, we propose Skip DETR, which improves the feature fusion between different network layers through skip connection and the introduction of spatial pyramid pooling layers so as to improve the detection results of small objects. We performed experiments on Forestry Pest Datasets, and the experimental results showed significant AP improvements in our method. When the value of IoU is 0.5, our method is 7.7% higher than the baseline and 6.1% higher than the detection result of small objects. Experimental results show that the application of skip connection and spatial pyramid pooling layer in the detection framework can effectively improve the effect of small-sample obiect detection.

1 Introduction

Object detection is one of the more important branches in the field of computer vision Zaidi et al. (2022), and it has been widely used in agricultural pest detection, crop condition detection, crop yield prediction, and other fields. In recent years, with the vigorous development of deep convolutional neural networks, the accuracy and performance of object detection tasks have been greatly improved. The identification and detection of forest pests provide a strong guarantee for crop yield growth and the agricultural economy Ngugi et al. (2021). However, compared with tasks such as autonomous driving and intelligent monitoring, forestry pest detection still has the following challenges: (1) There are fewer publicly available datasets for forest pests; (2) The detection targets are mostly small targets Huang et al. (2022).

At present, there are few studies on forest pest detection, and forestry pest datasets are relatively lacking. Most of the proposed solutions rely on traditional CNN structures, such as ResNet, GoogleNet, VGG, etc. The root cause of this problem is the lack of a large enough dataset of forestry pests to train specific CNN structures Ngugi et al. (2021). These traditional architectures have high computational requirements for pest identification tasks and require high-resolution image features in processing small target images, which will bring high computational complexity. Therefore, Carion et al. (2020) proposed an end-to-end object detection model (DETR) based on Transformers and achieved competitive results. The main hurdle in forestry pest datasets is that the only large, freely available datasets are the PlantVillage dataset and the Plant Disease Symptom Image Database (PDDB). The recently published forestry pest dataset solves the above problems well. The dataset contains 7163 images and 31 species of forestry pests Liu et al. (2022a).

Current mainstream non-end-to-end object detection frameworks include one-stage and two-stage detectors. They may face a huge amount of computation, which can greatly improve model training time and performance Sun et al. (2021a). The DETR architecture Carion et al. (2020), as an end-to-end object detection framework, has a simple structure and does not require specialized libraries. This means faster setup of deployments or downstream tasks on other computers. At the same time, experiments show that DETR also has high performance when migrating to other tasks, such as panoramic segmentation.

However, DETR, a method that uses object queries matching, usually requires high costs when performing intensive detection. There are many ways to improve DETR, such as the way Deformable DETR uses Deformable’s structure and Multi-Scale Zhu et al. (2020). It greatly reduces the training cost of the model while improving the performance of the model. At the same time, due to the lack of image priori and multi-scale fusion mechanisms of DETR Liu et al. (2022b), although recent DETR-based models have achieved significant performance, DETR lacks multi-scale features compared to classical object detection models, which are critical for small object detection.

As a classical method to improve the structure of deep neural networks, skip connection He et al. (2016) has been applied in classical network structures such as U-Net, ResNet, and DenseNet, which plays a role in improving the accuracy of image segmentation and improving the utilization rate of feature information in each layer of the network. Therefore, this paper introduces skip connection and spatial pyramid pooling layers He et al. (2015) to enhance the extraction and fusion of image features by the model and enhance the model’s learning of small object objects. First, we link the backbone network output and encoder output with skip connection. Then, in the DETR decoder input, 100 randomly initialized object queries are included, but random initialization also makes image feature learning slow Chen et al. (2022). So we use a three-layer spatial pyramid pooling layer to transform the output of the backbone network into features of the same size as the object queries, and finally perform skip connection to improve the initialization process of the object queries.

With our improved method, we have improved the accuracy of small target detection. At the same time, compared with the initial DETR model, our model has achieved competitive results. On Forestry Pest Datasets, at IoU=0.5 and IoU=0.75, our method achieves an absolute gain of 7.7% AP and 6.0% AP on the DETR baseline. For small object detection results, our method achieves a gain of 6.1% AP over the DETR baseline.

The contribution of this work are summarized as follows:

1) We propose a model called Skip DETR, which uses skip connection to enhance the extraction of features of small sample images by the DETR model.

2) We introduce the spatial pyramid pooling layer, improve the object queries initialization method, and make the model converge faster.

3) We conduct extensive experiments on forestry pest datasets. Experimental results show that the application of skip connection and spatial pyramid pooling layer in the detection framework can effectively improve the effect of small-sample object detection.

2 Related works

In this section, we will introduce the relevant solutions for insufficient multi-scale feature fusion and the research status of small target detection, identify and review existing forest pest datasets and related detection methods.

2.1 A solution to insufficient multi-scale feature fusion

The DETR architecture, as an end-to-end object detection framework, has a simple structure and does not require specialized libraries. However, the DETR model does not contain FPN, resulting in high computational complexity and insufficient feature fusion when processing high-resolution image features. However, DETR requires high-resolution image features when processing small target images, which brings high computational complexity. Therefore, it is not suitable to introduce FPN inside the DETR model, which ultimately leads to insufficient feature fusion. Without reintroducing multiscale feature fusion in the encoder, the accuracy of DETR cannot be further improved. Therefore, six Transformer encoder layers are included in the DETR encoder, which are stacked on top of the backbone network to improve the feature representation of its model.

At present, many improved models of DETR are trying to solve this problem. Deformable DETR combines DCN sparse sampling capabilities with transformer global relationship modeling capabilities by using the Deformable Attention module Zhu et al. (2020). Sun et al. (2021b) solved the cross-attention problem in DETR by proposing two schemes, TSP-FCOS and TSP-RCNN. The ViT-FRCNN model improves the structure of DETR by replacing the transformer with the backbone portion of FasterRCNN Beal et al. (2020). Since the introduction of FPN is not suitable inside DETR, the work in this paper is mainly to add skip connections and SPP networks outside DETR to enhance the fusion of its multi-scale features.

2.2 Public datasets of forest pests and pest object detection

The identification and detection of pests and diseases provide a strong guarantee for crop yield growth and agricultural economy in forestry pest control. Current forestry pest datasets can provide a wide variety of training samples for target detectors. Sun et al. (2018), as well as Hong et al. (2021), used pheromone traps to collect datasets that created forestry pests, but the datasets they created were only able to handle specific species of forestry pests. Chen et al. (2019) also created a dataset of forestry pests, but their main purpose was to study the classification of pests. Baidu has also published a dataset of forestry pests, but it was collected in a lab-built environment. Therefore, finding a public and suitable forestry pest dataset is difficult. However, the forestry pest dataset recently published by Liu et al. (2022a), which contains 31 pests and more than 7,000 images, lays a good foundation for the training of target detectors in this field.

Early pest and disease object detection was largely based on machine learning techniques. Le-Qing and Zhen (2012) tested 10 pests on a dataset of 579 samples using local average color features and SVMs.

Zhang et al. (2013) proposed a field pest identification system, and the dataset they used included about 270 training samples. Ebrahimi et al. (2017) used an SVM method with differential kernel functions for parasite classification and thrips detection. These early pest detection methods have yielded good results. However, their detection performance depends on the performance of the manual feature extractor and the chosen classifier.

With the development of image technology, convolutional neural networks have achieved obvious advantages in complex object detection, segmentation and classification by virtue of their strong image feature learning ability. Selvaraj et al. (2019) constructed an AI-based banana pest detection system based on deep convolutional neural network (DCNN). Liu and Wang (2020) constructed a tomato pest dataset and improved the YOLOV3 model to detect tomato pests and diseases based on this dataset. Zhu et al. (2021) improved the YOLOV3 model for the detection of black rot in grape leaves using super-resolution image enhancement.

In summary, although convolutional neural network (CNN)-based pest detection can improve the performance of pest detection, it has the advantage of avoiding the early limitations of the model. However, the fly in the ointment is that the vast majority of object detection architectures have manually designed components that have an impact on the performance of the model. Recently, the end-to-end object detection model (DETR) based on Transformers proposed by Carion et al. (2020) can avoid the above problems well and achieve competitive results.

2.3 Status of small target detection

Small target detection plays an important role in forestry pest detection, crop status detection, crop yield prediction and other scenarios Du et al. (2022). Small target detection has the characteristics of small coverage image area, few effective features for object detection, and commonly used object detectors are insensitive to small targets. At present, for small target detection, MR-CNN adopts multi-scale feature fusion Gidaris and Komodakis (2015), ContextNet uses context information to improve R-CNN Han et al. (2020), JCS-Net adopts image super-resolution and other methods Pang et al. (2019), which has been improved in the detection results of small targets.

However, most studies did not work on forest pest datasets, and there are still some gaps in the detection of small targets in the field of forest pest identification.

3 Framework of Skip DETR model

In this section, we will introduce the model structure of Skip DETR and elaborate on the structure of the improved components and how they provide gain to the model.

3.1 The structure of Skip DETR

Skip DETR is an improved end-to-end object detection framework based on DETR, which mainly includes three parts: backbone, transformer-based encoder-decoder structure, and sequence prediction architecture. At the same time, we add a spatial pyramid pooling layer and a deep separable convolutional layer outside the DETR model, and enhance the fusion of contextual feature information through skip connections.

When the image is input to the model, it will first be processed by the CNN to obtain the feature matrix of the current image. Then the feature matrix will be straightened and added to the position encoding, and passed into the encoder to learn the global information of the image, and the straightened feature matrix will be further extracted by deep separable convolutional layers. The results of the subsequent deep separable convolutional layer processing will be residually connected to the Encoder output on the one hand, and a three-layer spatial pyramid pooling layer and connected to the object query as the input of the Decoder on the other hand. Finally, it is decoded by Decoder and passed to FFN for image prediction. We will show the structure of our Skip DETR in Figure 1.

FIGURE 1
www.frontiersin.org

Figure 1 The structure of Skip DETR. Our work is mainly carried out outside the DETR model. We introduce a deep separable convolutional layer and a spatial pyramid pooling layer, and skip connection with the encoder output and object queries.

3.2 Applying skip connections between different layers of DETR

The basic idea of skip connection is to express the output as a linear superposition of a nonlinear transformation of the input and output He et al. (2016). After the skip connection, the amount of information describing the characteristics of the image increases, but the feature dimension describing the image does not change. Because the amount of information increases in each dimension, it is beneficial for the final image prediction.

When we use a non-linear change function to describe the input and output of a network, that is, the input is x and the output is F(x), F usually includes operations such as convolution and activation. When we add an input to the output of a function, although we can still use G(x) to describe the relationship between input and output, G(x) can be explicitly split into linear overlays of F(x) and X.

Srivastava et al. (2015) proposed the residual structure for the first time, which is derived from the control gate idea of LSTM. The initial residual formula is shown in formula 1.

y=H(x,WH )·T(x,WT )+x·(1T(x,WT ))(1)

However, because Formula 1 is too complex, He et al. (2016) simplified the formula, which is shown in formula 2.

y=H(x,WH )+X(2)

Inspired by the residual structure, this paper applies the residual structure to the DETR network. On the one hand, we link the backbone network output with the encoder output to enhance the learning of small objects by fusing image features. In order to reduce the number of model parameters and operation costs,

we introduce a depthwise separable convolutional layer after the output of the backbone network. We set the convolution kernel size to 1, the stride to 1, the depthwise part group to 256, and the pointwise part group to 1. We will show the structure of this part of the component in Figure 2.

FIGURE 2
www.frontiersin.org

Figure 2 Depthwise Separable Convolution and residual structure diagram. Depthwise Separable Convolution is divided into two parts: Depthwise Convolution and Pointwise Convolution. We set the relevant parameters according to the input image features. It is worth noting that the convolution kernel size of both parts is 1×1.

On the other hand, the decoder input of DETR contains 100 randomly initialized object queries, but random initialization also makes the model converge slowly. So we want to link the backbone network output with the object query to speed up the convergence of the model by improving the initialization of the object query. However, since the backbone network output feature map size is variable, it is not possible to unify the output feature map size using a common convolution layer. This paper solves this problem by introducing a spatial pyramid pooling layer, which will be described in the next section.

3.3 Add a spatial pyramid pooling layer

The essence of the pooling layer of the spatial pyramid is the multi-layer maximum pooling layer, which generates a fixed-size output for feature maps (n × n) of different sizes (α × α). The spatial pyramid pooling layer automatically adjusts the size of the sliding window win and the step size str according to different input sizes, using Equation 3 and Equation 4. In this paper, the output results of deeply separable convolutional layers are processed by spatial pyramid pooling layers and residually connected with object queries in the DETR model. This component changes the initial state of the object queries, providing the model with a priori information that can learn the key features of the image, thereby shortening the convergence process of the model. We will show the structure of this component in Figure 3.

FIGURE 3
www.frontiersin.org

Figure 3 Spatial pyramid pooling layer structure diagram. We design a three-layer spatial pyramid pooling layer, decompose the feature map into 2×2, 4×4, 8×8 sizes, and at the same time stitch the output of the second layer twice when unfolding and stitching. Finally, we connect the output with the object queries.

win=ceil(α/n)(3)
str=floor(α/n)(4)

Many CNN models have certain requirements on the size of the input image Long et al. (2015), but the feature extraction network (such as convolution layer, activation function layer, pooling layer) part of the model head has no requirements on the input image, which can be simply understood as the feature extraction network knowledge reduces the image by a fixed multiple. However, the full connection layer at the end of the model has strict requirements on the input dimension. Therefore, limiting the image size of the input CNN model is to meet the requirements of the full connection layer.

In the decoder input of the DETR, 100 randomly initialized object queries are included, and the size of each object query is 100 × 256. However, the size of the feature map output by the backbone network cannot be guaranteed to be the same as the size of object query. Therefore, we first pass the output of the backbone network through a Depthwise Separable Convolutional layer (the convolutional layer is described in Section 3.1), and then pass through a three-layer spatial pyramid pooling layer, so that the final output can meet the skip connection condition with the object query.

In the process of using the pooling layer of the spatial pyramid, we also hope to integrate the outputs of different pooling layers. Therefore, we integrate the output of the second pooling layer with the output of the third pooling layer, hoping to further enhance the model’s ability to extract features.

4 Experiment

We show that in the quantitative assessment of Forestry Pest Datasets, our improved DETR has achieved competitive results compared with the baseline of the DETR. Then we carried out a detailed ablation study and gave qualitative results.

4.1 Parameters of model training

The experiments used in this paper is Python 3.9, Torch 1.11, CUDA 11.4. Other hardware information used in the experiment is shown in Table 1. The main parameters of the Skip DETR model are shown in Table 2. DETR requires about 500 epochs on the COCO dataset to converge, considering the detection accuracy and training time on the Forestry Pest dataset, we choose 300 epochs. At the same time, in order to ensure the consistency of the training cycle of the control experiment, we designed the same epoch for other experiments. If there is no special declaration, all other parameters are consistent with the DETR.

TABLE 1
www.frontiersin.org

Table 1 Configuration of experimental environment.

TABLE 2
www.frontiersin.org

Table 2 Model parameter settings of Skip DETR.

4.2 Dataset

At present, there are many datasets related to the Forestry Pest identification task (such as Hong et al. (2021); Chen et al. (2019); Sun et al. (2018)), but they have problems such as few pest species, being unable to apply to the actual scene, and the data set is not open to the public. However, the recently published Forestry Pest dataset has solved the above problems well Liu et al. (2022a). The dataset contains 7163 images and 31 forest pests. The dataset is derived from Liu et al. (2022a), and the types and quantities of forest pests in the dataset are shown in Table 3.

TABLE 3
www.frontiersin.org

Table 3 Details of the types and quantities of forest pests in the dataset.

Therefore, we use the dataset of Liu et al. (2022a) for training. In order to ensure the training results, the Forestry Pest dataset is randomly divided according to the following proportion: (Train: Val=9:1): Test=9:1. That is, 5801 training images, 645 verification images and 717 test images for target detection tasks are included after division.

4.3 Evaluation metrics

In this paper, we use mAP and AR as experimental evaluation indicators, which are widely used in the field of object detection. We will give the calculation method of mAP and AR.

Precision=TPTP+FP(5)
Recall=TPTP+FN(6)
mAPα=1Nn=1NAPαn(7)

where TP for positive samples is predicted as positive class, FP is negative samples are predicted as positive class, and FN is positive samples are predicted as negative class. AP is the average accuracy, which is simply the average of the Precision value on the PR curve, and mAPα represents the AP measurement at different IoU thresholds.

In the COCO dataset, objects with a pixel area less than 32 × 32 are regarded as small objects, pixel faces and objects larger than 96 × 96 are regarded as large objects, and pixel faces and objects between 32 × 32 and 96 × 96 are regarded as medium objects.

4.4 Experimental results

Skip DETR is an improvement based on the DETR model and is mainly designed for small target detection.

We try to improve the performance improvement of DETR on small target detection with our method. The average accuracy of the Skip DETR model at different IoU Rezatofighi et al. (2019) thresholds, the results are shown in Table 4.

TABLE 4
www.frontiersin.org

Table 4 The average accuracy of the Skip DETR model at different IoU thresholds.

From the experimental results in Table 4, it can be seen that after training with 300 epochs, the Skip DETR model has better accuracy than the DETR model on the forest pest dataset. When IoU=0.5, The 200th epoch result of Skip DETR is even higher than the 300th epoch result of the DETR model. At the same time, the results on other evaluation indicators are also due to the DETR model. This shows that the improvement method we use helps to improve the accuracy of the model.

Another drawback of DETR is its poor performance in detecting small objects. To verify whether our model helps improve the accuracy of small object detection, we compare the detection accuracy of Skip DETR and DETR at different scales. The results are shown in Table 5.

TABLE 5
www.frontiersin.org

Table 5 The detection accuracy of Skip DETR and DETR at different scales.

As can be seen from the results in Table 4, our model is a significant improvement in the detection of small objects. After 300 epochs, compared with DETR, the accuracy of skip DETR in small object detection is improved by 6.1% AP, the medium object detection accuracy is improved by 11.3% AP, and the detection accuracy of large objects is improved by 8.8% AP.

Recall is often used to assess detector coverage of all objects to be inspected Buckland and Gey (1994).Therefore, we compared the recall of Skip DETR and DETR at different training stages and scales in Table 6. We selected 100 subjects to test the average recall, and the final result showed that Skip DETR can predict positive samples more accurately.

TABLE 6
www.frontiersin.org

Table 6 The recall of Skip DETR and DETR at different training stages and scales.

Our improved DETR model has obtained good results, especially greatly improved APS, APMand APL.

Finally, we used the original DETR model and the improved DETR model to measure GFLOPS Goldberg (1991) and params. Through experiments we find that the improvement method proposed in this paper is completely negligible in terms of consumption of computing resources.

During model training, training loss Ru et al. (2020), box loss, and classification error rate Kim (2009) are common metrics to measure model performance. In Figure 4 we show the comparison results of the Skip DETR and DETR models on the above evaluation indicators. From Figure 4, it can be seen that the loss error of Skip DETR in the initial training phase is lower than that of the DETR model. As training progresses, Skip DETR’s loss and classification error rate decreases faster than DETR. This indicates the effectiveness of adding skip connection and spatial pyramid pooling layers. In addition, the training process of Skip DETR is smoother, which is easier to train than the DETR model.

FIGURE 4
www.frontiersin.org

Figure 4 Comparison of common model evaluation indicators. (A) Training loss comparison; (B) Box loss comparison; (C) Class error comparison.

Furthermore, in Table 7, we compare our Skip DETR with several different object detection models on Forestry Pest Datasets. We report the detection results of each model at different IoU thresholds and scales. For fair comparison, we used the same model parameters and trained the same epochs. We show that Skip DETR outperforms DETR and achieves competitive results compared with other object detection models and improved models based on DETR.

TABLE 7
www.frontiersin.org

Table 7 Ablation study for each part of the contribution to DETR.

4.5 Ablation experiments

In this section, we conducted several ablation experiments to help us understand the contribution of each improved method to the final performance. As shown in Table 8, both improvement methods improve the performance of DETR.

TABLE 8
www.frontiersin.org

Table 8 Ablation study for each part of the contribution to DETR.

It is worth noting that we first introduced the spatial pyramid pooling layer on the basis of DETR, and we found that although it provides performance improvement for DETR in the recognition of small objects, it will reduce the performance of other indicators. Therefore, we introduced the skip connection on the basis of the first improvement, and experimented with the skip connection and the spatial pyramid pooling layer as a whole module, and finally obtained the performance improvement on all indicators while the number of model parameters remained basically unchanged.

In general, in the field of forestry pest detection and small target detection, Skip DETR adds skip connections and spatial pyramid pooling layer so that our model can make full use of the image information in the feature map at various scales, and making the model more sensitive to small targets. At the same time, the spatial pyramid pooling layer can change the initialization mode of object queries, making the convergence of Skip DETR models faster and easier to train. Without changing the number of model parameters too much, Skip DETR has achieved competitive results on multiple evaluation indicators. And compared to several other different object detection models, Skip DETR also achieves better results.

5 Conclusion

In this work, we propose a model called Skip DETR, which uses skip connection to enhance the extraction of image features from small samples by the DETR model. At the same time, we introduce the spatial pyramid pooling layer, improve the object query initialization method, and make the model converge faster. Finally, we conduct extensive experiments on forestry pest datasets. Experimental results show that the application of skip connection and spatial pyramid pooling layer in the detection framework can effectively improve the effect of small-sample object detection.

Although Skip DETR achieved good results, our study still faced the problem of small data pools. At the same time, in order to improve the detection accuracy, we will continue to improve the Skip DETR model.

Data availability statement

The original contributions presented in the study are included in the article/supplementary files, further inquiries can be directed to the corresponding author.

Author contributions

BL and YJ conducted experiments, analyzed data, and wrote manuscripts. LL provided a dataset of forestry pests. YD and SS designed the research and revised the manuscript. All authors contributed to the article and approved the submitted version.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Beal, J., Kim, E., Tzeng, E., Park, D. H., Zhai, A., Kislyuk, D. (2020). Toward transformer-based object detection. arXiv. preprint. arXiv:2012.09958.

Google Scholar

Buckland, M., Gey, F. (1994). The relationship between recall and precision. J. Am. Soc. Inf. Sci. 45, 12–19. doi: 10.1002/(SICI)1097-4571(199401)45:1<12::AID-ASI2>3.0.CO;2-L

CrossRef Full Text | Google Scholar

Carion, N., Massa, F., Synnaeve, G., Usunier, N., Kirillov, A., Zagoruyko, S. (2020). “End-to-end object detection with transformers,” in Computer Vision–ECCV 2020: 16th European Conference, August 23–28, 2020. 213–229 (Glasgow, UK: Springer).

Google Scholar

Chen, J., Chen, L., Wang, S., Zhao, H., Wen, C. (2019). Pest image recognition of garden based on improved residual network. Trans. Chin. Soc Agric. Machi. 50, 187–195.

Google Scholar

Chen, X., Wei, F., Zeng, G., Wang, J. (2022). Conditional detr v2: Efficient detection transformer with box queries. arXiv. preprint. arXiv:2207.08914.

Google Scholar

Du, J., Liu, L., Li, R., Jiao, L., Xie, C., Wang, R. (2022). Towards densely clustered tiny pest detection in the wild environment. Neurocomputing 490, 400–412. doi: 10.1016/j.neucom.2021.12.012

CrossRef Full Text | Google Scholar

Ebrahimi, M., Khoshtaghaza, M. H., Minaei, S., Jamshidi, B. (2017). Vision-based pest detection based on svm classification method. Comput. Electron. Agric. 137, 52–58. doi: 10.1016/j.compag.2017.03.016

CrossRef Full Text | Google Scholar

Gidaris, S., Komodakis, N. (2015). “Object detection via a multi-region and semantic segmentation-aware cnn model,” in Proceedings of the IEEE international conference on computer vision (Santiago, Chile) 1134–1142.

Google Scholar

Goldberg, D. (1991). What every computer scientist should know about floating-point arithmetic. ACM Computing. Surveys. (CSUR). 23, 5–48. doi: 10.1145/103162.103163

CrossRef Full Text | Google Scholar

Han, W., Zhang, Z., Zhang, Y., Yu, J., Chiu, C.-C., Qin, J., et al. (2020). Contextnet: Improving convolutional neural networks for automatic speech recognition with global context. arXiv. preprint. arXiv:2005.03191.

Google Scholar

He, K., Zhang, X., Ren, S., Sun, J. (2015). Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 37, 1904–1916. doi: 10.1109/TPAMI.2015.2389824

PubMed Abstract | CrossRef Full Text | Google Scholar

He, K., Zhang, X., Ren, S., Sun, J. (2016). “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition (Las Vegas, NV, USA) 770–778.

Google Scholar

Hong, S.-J., Nam, I., Kim, S.-Y., Kim, E., Lee, C.-H., Ahn, S., et al. (2021). Automatic pest counting from pheromone trap images using deep learning object detectors for matsucoccus thunbergianae monitoring. Insects 12, 342. doi: 10.3390/insects12040342

PubMed Abstract | CrossRef Full Text | Google Scholar

Huang, J., Huang, Y., Huang, H., Zhu, W., Zhang, J., Zhou, X. (2022). An improved yolox algorithm for forest insect pest detection. Comput. Intell. Neurosci. 2022. doi: 10.1155/2022/5787554

CrossRef Full Text | Google Scholar

Kim, J.-H. (2009). Estimating classification error rate: Repeated cross-validation, repeated hold-out and bootstrap. Comput. Stat Data Anal. 53, 3735–3745. doi: 10.1016/j.csda.2009.04.009

CrossRef Full Text | Google Scholar

Le-Qing, Z., Zhen, Z. (2012). Automatic insect classification based on local mean colour feature and supported vector machines. Oriental. Insects. 46, 260–269. doi: 10.1080/00305316.2012.738142

CrossRef Full Text | Google Scholar

Liu, S., Li, F., Zhang, H., Yang, X., Qi, X., Su, H., et al. (2022b). Dab-detr: Dynamic anchor boxes are better queries for detr. arXiv. preprint. arXiv:2201.12329.

Google Scholar

Liu, B., Liu, L., Zhuo, R., Chen, W., Duan, R., Wang, G. (2022a). A dataset for forestry pest identification. Front. Plant Sci. 13. doi: 10.3389/fpls.2022.857104

CrossRef Full Text | Google Scholar

Liu, J., Wang, X. (2020). Tomato diseases and pests detection based on improved yolo v3 convolutional neural network. Front. Plant Sci. 11, 898. doi: 10.3389/fpls.2020.00898

PubMed Abstract | CrossRef Full Text | Google Scholar

Long, J., Shelhamer, E., Darrell, T. (2015). “Fully convolutional networks for semantic segmentation,” in Proceedings of the IEEE conference on computer vision and pattern recognition (Boston, MA, USA) 3431–3440.

Google Scholar

Ngugi, L. C., Abelwahab, M., Abo-Zahhad, M. (2021). Recent advances in image processing techniques for automated leaf pest and disease recognition–a review. Inf. Process. Agric. 8, 27–51. doi: 10.1016/j.inpa.2020.04.004

CrossRef Full Text | Google Scholar

Pang, Y., Cao, J., Wang, J., Han, J. (2019). Jcs-net: Joint classification and super-resolution network for small-scale pedestrian detection in surveillance images. IEEE Trans. Inf. Forensics. Secur. 14, 3322–3331. doi: 10.1109/TIFS.2019.2916592

CrossRef Full Text | Google Scholar

Rezatofighi, H., Tsoi, N., Gwak, J., Sadeghian, A., Reid, I., Savarese, S. (2019). “Generalized intersection over union: A metric and a loss for bounding box regression,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (Seoul, South Korea) 658–666.

Google Scholar

Ru, B., Lyle, C., Schut, L., van der Wilk, M., Gal, Y. (2020). Revisiting the train loss: an efficient performance estimator for neural architecture search. stat 1050, 8.

Google Scholar

Selvaraj, M. G., Vergara, A., Ruiz, H., Safari, N., Elayabalan, S., Ocimati, W., et al. (2019). Ai-powered banana diseases and pest detection. Plant Methods 15, 1–11. doi: 10.1186/s13007-019-0475-z

PubMed Abstract | CrossRef Full Text | Google Scholar

Srivastava, R. K., Greff, K., Schmidhuber, J. (2015). Highway networks. arXiv. preprint. arXiv:1505.00387.

Google Scholar

Sun, Z., Cao, S., Yang, Y., Kitani, K. M. (2021b). “Rethinking transformer-based set prediction for object detection,” in Proceedings of the IEEE/CVF international conference on computer vision (Nashville, TN, USA) 3611–3620.

Google Scholar

Sun, Y., Liu, X., Yuan, M., Ren, L., Wang, J., Chen, Z. (2018). Automatic in-trap pest detection using deep learning for pheromone-based dendroctonus valens monitoring. Biosyst. Eng. 176, 140–150. doi: 10.1016/j.biosystemseng.2018.10.012

CrossRef Full Text | Google Scholar

Sun, P., Zhang, R., Jiang, Y., Kong, T., Xu, C., Zhan, W., et al. (2021a). Sparse r-cnn: End-to-end object detection with learnable proposals. In. Proc. IEEE/CVF. Conf. Comput. Vision Pattern Recognit. 14454–14463. doi: 10.1109/CVPR46437.2021.01422

CrossRef Full Text | Google Scholar

Zaidi, S. S. A., Ansari, M. S., Aslam, A., Kanwal, N., Asghar, M., Lee, B. (2022). A survey of modern deep learning based object detection models. Digital. Signal Process. 103514.

Google Scholar

Zhang, H. T., Hu, Y. X., Zhang, H. Y. (2013). Extraction and classifier design for image recognition of insect pests on field crops. Adv. Mater. Res. 756, 4063–4067. doi: 10.4028/www.scientific.net/AMR.756-759.4063

CrossRef Full Text | Google Scholar

Zhu, J., Cheng, M., Wang, Q., Yuan, H., Cai, Z. (2021). Grape leaf black rot detection based on super-resolution image enhancement and deep learning. Front. Plant Sci. 12, 695749. doi: 10.3389/fpls.2021.695749

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhu, X., Su, W., Lu, L., Li, B., Wang, X., Dai, J. (2020). Deformable detr: Deformable transformers for end-to-end object detection. arXiv. preprint. arXiv:2010.04159.

Google Scholar

Keywords: object detection, forestry pest detection, DETR, Skip connection, small object detection

Citation: Liu B, Jia Y, Liu L, Dang Y and Song S (2023) Skip DETR: end-to-end Skip connection model for small object detection in forestry pest dataset. Front. Plant Sci. 14:1219474. doi: 10.3389/fpls.2023.1219474

Received: 09 May 2023; Accepted: 25 July 2023;
Published: 15 August 2023.

Edited by:

Liangxiu Han, Manchester Metropolitan University, United Kingdom

Reviewed by:

Pankaj Kumar Tyagi, Noida Institute of Engineering and Technology (NIET), India
Jun Liu, Weifang University of Science and Technology, China

Copyright © 2023 Liu, Jia, Liu, Dang and Song. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Shinan Song, songshinan@ccut.edu.cn

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.