Skip to main content

ORIGINAL RESEARCH article

Front. Bioeng. Biotechnol., 21 July 2023
Sec. Biosensors and Biomolecular Electronics

Image recognition of traditional Chinese medicine based on deep learning

Junfeng MiaoJunfeng Miao1Yanan Huang
Yanan Huang2*Zhaoshun WangZhaoshun Wang1Zeqing WuZeqing Wu3Jianhui LvJianhui Lv4
  • 1School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, China
  • 2Business School, Ezhou Vocational University, Ezhou, Hubei, China
  • 3School of Pharmacy, Xinxiang Medical University, Xinxiang, China
  • 4Pengcheng Laboratory, Shenzhen, China

Chinese herbal medicine is an essential part of traditional Chinese medicine and herbalism, and has important significance in the treatment combined with modern medicine. The correct use of Chinese herbal medicine, including identification and classification, is crucial to the life safety of patients. Recently, deep learning has achieved advanced performance in image classification, and researchers have applied this technology to carry out classification work on traditional Chinese medicine and its products. Therefore, this paper uses the improved ConvNeXt network to extract features and classify traditional Chinese medicine. Its structure is to fuse ConvNeXt with ACMix network to improve the performance of ConvNeXt feature extraction. Through using data processing and data augmentation techniques, the sample size is indirectly expanded, the generalization ability is enhanced, and the feature extraction ability is improved. A traditional Chinese medicine classification model is established, and the good recognition results are achieved. Finally, the effectiveness of traditional Chinese medicine identification is verified through the established classification model, and different depth of network models are compared to improve the efficiency and accuracy of the model.

1 Introduction

Chinese medicinal materials are particularly important in the growth of traditional Chinese medicine, which plays an irreplaceable role in treating diseases (Chen and He, 2022). It is also an important component of the most complete traditional medical system in the world today (Xing et al., 2020). Traditional Chinese medicine is also a continuous understanding of medicinal plants that grow in nature. After so many years of development, there are more and more varieties of traditional Chinese medicine, most of which are herbaceous plants, as well as some animal products, minerals, etc (Chen et al., 2021). However, there are also many different types of sub varieties under the same variety, so effective methods are needed to identify the varieties and quality of traditional Chinese medicine. Therefore, the identification and classification of traditional Chinese medicine has always been a hot issue that urgently needs further improvement (Hao et al., 2021). The traditional methods for evaluating the quality of traditional Chinese medicinal materials mainly rely on empirical identification, which can effectively reflect the type and quality of Chinese medicinal materials (Zhou et al., 2021). However, traditional empirical identification is achieved through the subjective feelings and relevant knowledge reserves of professionals (Wang et al., 2020). However, in practical applications, this manual recognition method is cumbersome, inefficient, and the judgment criteria often have subjectivity, posing a risk of misjudgment, affecting the safety and effectiveness of clinical medication, and seriously hindering the healthy development of the traditional Chinese medicine industry (Chu et al., 2022). Therefore, exploring an intelligent, efficient, and high-precision method for identifying traditional Chinese medicine is of great practical significance.

With the rapid development of internet technology and supercomputing, artificial intelligence has gradually spread throughout various fields of human production and life (Wang et al., 2019). In the era of intelligent development, humans are constantly trying to obtain more information from images in better ways. By using computer image recognition technology, a traditional Chinese medicine classification and recognition system is established to efficiently classify and organize a large number of images (Xu et al., 2021). To a certain extent, it can help professionals identify their types, reduce workload, improve work efficiency and recognition accuracy. The deep learning methods in traditional image recognition mainly extract low-level information such as color, texture, shape, etc. It relies on artificially designed features and is difficult to extract high-level semantic features of the image (Zhao et al., 2022). In recent years, deep learning has sparked a wave of enthusiasm in the academic community, widely applied in image recognition and achieved breakthrough results (Wang et al., 2023). It can effectively map low-level features to high-level fields, obtain more essential feature representations, and ensure higher recognition rates while providing more convenient operations. Deep learning is a reliable method to improve the efficiency and accuracy of identifying traditional Chinese medicine types (Han et al., 2022). Therefore, this article uses convolutional neural networks to complete the identification and classification of traditional Chinese medicine.

The main work and structure of this article are as follows.

(1) This article studies the development of traditional Chinese medicine recognition, identifies the targets that need to be recognized, and constructs a dataset of traditional Chinese medicine images.

(2) This article preprocesses images of traditional Chinese medicine. By normalizing, graying, and image denoising operations, the quality of traditional Chinese medicine images has been improved. Then, in order to solve the problem of a small number of image datasets for traditional Chinese medicine, data augmentation techniques are used to ensure the adequacy of the image dataset.

(3) This article uses an improved ConvNeXt network for feature extraction and classification of traditional Chinese medicine. This method adds a Stacked ACMix module to the second ConvNeXt Block layer inside the ConvNeXt network to ensure the adequacy of low-dimensional feature extraction in the image. Then, a Stacked ACMix module is added to the last ConvNeXt Block layer inside the ConvNeXt network to ensure high-dimensional feature extraction and achieve sufficient information fusion. In the linear layer of the Head, the Stacked FFN network is added to ensure the specificity of its application in traditional Chinese medicine recognition. Finally, a comparative experiment is conducted on network models of different depths, and a network model with better classification performance is selected.

The remaining part of this article consists of four parts, and Section 2 is the related literature of this article. Section 3 provides a detailed introduction to Convolutional neural network, dataset for medicinal image, and recognition model. Section 4 analyzes the proposed algorithm and its effectiveness through experiments. Finally, the conclusion of this article is summarized.

2 Literature review

The combination of computer technology and traditional Chinese medicine identification and recognition has begun and in traditional Chinese medicine detection has received some research. Wu et al. (Wu et al., 2016) obtained the odor and color characteristics of traditional Chinese medicine using electronic nose and computer vision, and then used BP and SVM methods for classification and recognition; Hussein et al. (Hussein et al., 2020) identified forest plants using discriminant analysis, random forests, and support vector machines, showing the feasibility of using extracted traits to identify plants. The method described above has achieved good results to some extent, but has the following limitations: shallow features are the feature information that directly comes from image pixels without high-level semantics, and are easily affected by the detection environment. In practical applications, recognition reliability is poor (Sun and Qian, 2016). At the same time, feature extraction algorithms are established for visual shapes, colors, and textures, and then fused. The algorithm is complex and fails to consider the correlation between shapes, colors, and textures, resulting in poor results and reduced recognition efficiency. Therefore, the detection technology of traditional Chinese medicine based on image processing and pattern recognition is greatly limited in practice, and a new identification method is urgently needed.

Recently, great advances have been achieved in artificial intelligence, image recognition, and other technologies, greatly promoting related technologies. In order to develop a fast, automatic, and accurate image recognition system, convolutional neural network (CNN) has received widespread attention (Artzai et al., 2019). In particular, CNN can complete the extraction of low-level features to high-level semantics (Kattenborn et al., 2019), significantly increasing the accuracy of identification and identification in various fields. Park et al. (Park et al., 2017) proposed a player evaluation model based on deep learning to analyze the impact on baseball leagues. Batchuluun et al. (Batchuluun et al., 2018) used CNN models to recognize human bodies in motion. Hansen et al. (Hansena et al., 2018) implemented the recognition of individual pigs using deep learning methods, with an accuracy rate of 96.7%. Fricker et al. (Fricker et al., 2019) proposed a hyperspectral tree image recognition method using CNN models. Altuntas et al. (Altuntaş et al., 2019) used CNN models to automatically identify haploid and diploid maize seeds. Zhang et al. (Zhang et al., 2019) proposed a network model based on void convolution for identifying six common cucumber plant diseases. While related research on traditional Chinese medicine recognition and deep learning methods have also begun to be combined, Lee et al. (Lee et al., 2015) studied a deep learning method that used deconvolution networks (DN) to learn recognition features from 44 different plant images, proving that learning features through CNN was superior to traditional manual features. Hu et al. (Hu et al., 2018) suggested and obtained excellent classification results with a multi-scale fusion convolutional neural network for plant leaf identification. Zhu et al. (Zhu et al., 2018) improved the deep convolutional neural network by dividing the initial picture into smaller pictures and loading them into the network, achieving rapid and accurate classification of plant leaves. Park et al. (Park et al., 2021) designed a multi rate three-dimensional convolutional neural network. This method enhanced the image and classified it through a joint fusion classifier. The experimental results had high accuracy. Liu et al. (Liu et al., 2022a) designed a network for recognizing animal fur. This network was based on feature fusion and fully utilized the texture information and inverted feature information of fur images, improving accuracy on the dataset. Jeong et al. (Jeong et al., 2020) proposed a facial expression recognition method. This method used three-dimensional convolution to simultaneously extract spatial and temporal features. Then, these features were combined through a joint fusion classifier to complement each other. Kim et al. (Kim et al., 2019) proposed a new method for FER systems based on hierarchical deep learning. This method integrated features extracted from appearance networks with geometric features in a hierarchical structure and improves accuracy on the dataset.

3 Image recognition of traditional Chinese medicine based on deep learning

Deep learning technology has been gradually applied to Natural language processing, voice, image and other aspects (Wang et al., 2021), and has made breakthrough progress in the application field of artificial intelligence, as well as excellent achievements in many fields (Wang et al., 2022). In order to solve the image recognition of traditional Chinese medicine, deep learning technology is used to complete the recognition of traditional Chinese medicine, which not only effectively solves the problems of misjudgment and low efficiency caused by human factors, but also has high detection and recognition rate, convenient operation, and greatly saves time and labor costs. The main process is shown in Figure 1.

FIGURE 1
www.frontiersin.org

FIGURE 1. Maximum pooling and average pooling

3.1 Convolutional neural network

Convolutional neural network can learn the features of images and are widely used in computer vision tasks such as image classification and object detection (Namatēvs, 2017). In recent years, with the continuous deepening of research, the structure of convolutional neural networks has been continuously improved, resulting in many classic network structures (Hubel and Wiesel, 1962; Fukushima and Miyake, 1982). Convolutional neural network is mainly composed of convolution layers, pooling layer, and fully connectivity layer. By repeatedly stacking convolution and pooling layers, convolutional neural network can extract local and global feature information in images, and then complete tasks such as classification or regression through fully connectivity layer.The convolutional neural network (LeCun et al., 1998) has five hierarchical structures.

(1) Input layer

The input layer of convolutional neural network is used to input raw or preprocessed data into the network.

(2) Convolution layer

Convolution layer is a linear computational layer that uses a series of convolutional kernels and multi-channel input data for convolution. Convolution is a fundamental operation in analytical mathematics, in which all the functions used to perform operations on input data are called convolutional kernels. Convolutional operations refer to the process of the convolutional kernel making small weighted sums at various positions of input data in the form of sliding windows. In practical operations, the convolution layer will use different numbers of convolution cores for convolution operations, and the weight of each convolution core will always remain unchanged. This procedure could decrease the number of parameters, speed up neural network convergence, and shorten training time.

(3) Pooling layer

The pooling layer further processes feature maps and effectively reduces and filters the size and valuable feature information within them. Convolution layer and full connectivity layer are difficult to converge and compute feature maps that control size during the operation process, resulting in the appearance and extraction of some useless features. So during the process of changing image size, the pooling layer can effectively reduce the redundancy of features by filtering and processing useless features.

Pooling methods generally include maximum pooling and average pooling. The maximum pooling selection preserves the maximum value in the local area of the feature map as the result value after pooling. For mean pooling, the local average value of the feature map is selected as the pooling result. Figure 2 shows the pooling operation

(4) Full connectivity layer

FIGURE 2
www.frontiersin.org

FIGURE 2. Maximum pooling and average pooling.

The full connectivity layer is used in the network structure as a way to connect with various nodes in the previous layer using full connectivity. Its function is to transform the image feature maps output in the convolutional layer into vector form. At the same time, the dimension of the image feature map is transformed into a one-dimensional vector and mapped into the sample label space, which is then input into the classifier in the network model for recognition and classification (Szegedy et al., 2015). The parameters of the fully connectivity layer account for a significant proportion of all parameters in the convolutional neural network model, resulting in a large number of parameters that consume a significant amount of storage and computational resources. And it is easy to cause the problem of overfitting of network model, which has an important impact on the improvement of network training efficiency.

(5) Output layer

The output layer is used to output numerical values calculated by convolutional neural network. The form of output varies for different problems.

3.2 Construction of medicinal material image dataset

Due to the lack of publicly available datasets for traditional Chinese medicine, and the relatively high requirement for dataset size in deep learning classification tasks, only sufficient datasets can fully train and improve the model, obtaining excellent models with strong generalization ability (Krizhevsky et al., 2012). Therefore, the dataset selected in this article is shown in Figure 3, which mainly consists of six categories: angelica sinensis radix, citrus reticulatae pericarpium, angelica dahuricae radix, lilii bulbus, lonicerae japonicae flos, glycyrrhizae radix et rhizome. These data mainly come from filming, network crawling, and so on.

FIGURE 3
www.frontiersin.org

FIGURE 3. Image samples.

3.3 Image data preprocessing

Image samples are the data foundation for image recognition algorithm research. Compared to directly inputting the original image into the network for training, the preprocessed data samples are easier to train and have better training results (He et al., 2016a). In addition, a comprehensive database can help improve the generalization ability and robustness of the network. In deep learning, image quality affects recognition accuracy, and the number of images affects network generalization performance.

3.3.1 Image data normalization

In deep learning, the image is usually normalized before model training, and normalizing the image adjusts the size of the feature values to a similar range. If not normalized, the gradient value will also be larger when the feature value is larger, and smaller when the feature value is smaller (Simonyan and Zisserman, 2014). Therefore, in order to make the model training converge smoothly, the image is normalized and the feature values of different dimensions are adjusted to similar ranges[36]. In this paper, the linear function conversion method is selected for image normalization, because the linear function conversion method does not involve the calculation of distance measurement and covariance, which is relatively simple and suitable for image normalization in image processing. The formula is as follows.

b=aMax/MaxMin(1)

In the formula, b is the output of image pixel values, a is the input of image pixel values, and max and min are the maximum and minimum pixel values.

3.3.2 Grayscale image data

Image grayscale is a grayscale technique that processes the three channels of RGB with color images into a single channel grayscale image (Shotton et al., 2013). In the process of color image processing, it is necessary to sequentially process the three RGB colors, which not only increases the time cost but also increases the processing pressure on hardware devices. Therefore, in order to solve the problems caused by color image processing, it is necessary to perform grayscale preprocessing operations on color images. Single channel grayscale images also have the effect of reflecting the contrast between the chromaticity and brightness levels of the entire image. Each type of color image has its own morphological feature representation, and each color image adopts different grayscale processing methods. The grayscale processing methods for color images include component grayscale, maximum grayscale, average grayscale, and weighted average grayscale (Huang et al., 2021). Here, the weighted average method is applyied to grayize the medicinal image.

3.3.3 Image data denoising

During the processing of images, it is inevitable that they will be affected by various intensity signals, which will affect the quality of the image and disrupt the correlation between the content structure and pixels in the image, which is not conducive to further analysis of the image (Han et al., 2022). The goal of image denoising is to improve the quality of the specified image and solve the problem of image quality degradation caused by noise interference. This article selects median filtering to denoise Chinese herbal medicine images. Median filtering is a nonlinear image processing technique that preserves more details of the image and does not cause image blurring issues. And it does not replace the central target pixel with the average of the pixels in the template, but sorts all the pixels in the template, taking the median of the sorted template pixel sequence as the value of the target pixel. In addition, median filtering not only has a significant effect on eliminating isolated noise points, but also has a good removal effect on slightly dense noise points. And it also has a particularly good removal effect on salt and pepper noise. Overall, the denoising effect of median filtering is superior to other algorithms (Hao et al., 2022).

3.4 Data enhancement

The problem of image recognition based on deep learning usually requires large-scale training data samples, otherwise the problem of overfitting will occur because the model is too complex and the amount of data is too small. In fact, it is difficult for people to directly collect the amount of data that meets the requirements of deep neural network training. The collection and screening of training data is a very time-consuming and mental task, and the workload is extremely large. Therefore, this article adopts the method of image data enhancement to increase the number of image samples (Wang et al., 2023). First, image rotation and mirror symmetry are performed on the traditional Chinese medicine image. Image rotation refers to rotating all pixels of an image at an angle of 0-360 around the center of the image. Mirror symmetry refers to the exchange of all pixels in the image by using the vertical line in the image as the axis. Then, the image difference method is used to change the size of the original image to different sizes, and new images are randomly cropped on the scaled image (Wang et al., 2021).

3.5 Construction of traditional Chinese medicine image recognition model

Convolutional neural network has received widespread attention in the field of computer vision in international academia and industry in recent years due to their relatively fast development compared to other image processing methods, In the current development of cutting-edge technology and the resolution of research hotspots, convolutional neural network has become the most effective and practical tool for solving practical engineering problems in the field of computer vision. Over time, many mainstream convolutional neural network models have emerged, such as GoogLeNet (Szegedy et al., 2015), VGG (Simonyan and Zisserman, 2015), ResNet (He et al., 2016b) and ConvNeXt (Liu et al., 2022b). This article mainly builds the model based on ConvNeXt.

3.5.1 Building an identification model

The convolutional neural network algorithm has been the mainstream idea in the field of image recognition to solve the problem of image recognition, until the emergence of the Vision Transformer (Dosovitskiy et al., 2020) algorithm, which transferred the Transformer idea from the NLP field to the image field. Afterwards, a series of new algorithms such as ViT, Swin Transformer, and DeiT are born. With the use of Transformer algorithm in images, there is a growing trend in research to replace convolutional neural networks with Transformer algorithm. Until the birth of the ConvNeXt algorithm, it is once again demonstrated that convolutional neural networks are still an effective algorithm in the field of image recognition. This article is mainly based on the ConvNeXt algorithm and improves it. It mainly adds ACMix to the ConvNeXt algorithm, which includes convolution operations and Multi-Head Self Attention operations. It can extract the image features of traditional Chinese medicine through two different operations, and finally fuse the features to output the fused results. At the same time, it adds a stacked FFN structure to the Head layer of ConvNeXt to output the category of traditional Chinese medicine. The specific algorithm framework is shown in Figure 4.

FIGURE 4
www.frontiersin.org

FIGURE 4. Recognition model.

The traditional Chinese medicine recognition model constructed in this article has the following structure. The image size of the traditional Chinese medicine after image preprocessing is 224×224×3. It first passes through a two-dimensional convolutional layer with Kernel 4 and Strike 4, and then performs Layer Norm normalization on the convolutional images. Then the image will be compressed into a feature image of 56×56×96. The image passes through m ConNeXt Block layers again, with a Dimension of 96. The feature image output from the ConNeXt Block layer is downsampled, and the width and height of the downsampled image are reduced by half, resulting in a final feature image size of 28×28×96. Through the stacked ConvNeXt Block layer with a Dimension of 192, the output features have become 28×28×192. The image features output by the Stacked ConvNeXt Block layer are input into the Stacked ACMix module, which allows for varying degrees of extraction of features from the image through both Attention and Convolution operations. By integrating the advantages of two different frameworks, the extracted features can focus on both local features and global features through the Attention layer. The feature size output by ACMix is still 28×28×192. Later, the Downsampling of this feature becomes 14×14×192. The feature image of 14×14×192 is fed into the Stacked ConvNeXt Block layer with a Dimension of 384, and the output feature is 14×14×384. The feature map of 14×14×384 is downsampled again to become 7×7×384. The image features of 7×7×384 are processed through the Stacked ConvNeXt Block layer and output as a feature image of 7×7×768. The feature image is input into the Stacked ACMix. The output is 7×7×768. The output features use global average pooling again, and then pass through the Layer Norm layer. After normalization through the Layer Norm layer, it is inputted into the Stacked FFN network, and finally the classification head is output to complete the image recognition of traditional Chinese medicine.

3.5.2 Parameter setting

In the network designed in this article, the setting of relevant parameters has a significant impact on the accuracy of classification. The EPOCH hyperparameter in this paper is 100. The parameters of M,N and K are parameters that need to be optimized during operation. The different values of M, N and K directly determine the number of layers in the network. This article uses a pre-trained model. ConvNeXt has 5 models including ConvNeXt-nano model, ConvNeXt-Tiny model, ConvNeXt-Small model, ConvNeXt-Base model, and ConvNeXt-Large model. This article is an improvement on the ConvNeXt Tiny model. The values of K selected in this article are a set composed of 1,2 and 3. The initial value of learning rate γ is 0.0002. The maximum width of the network is 768.

4 Experiment and result analysis

4.1 Experimental environment and design

The experimental environment for this article is Linux operating system, Intel Core CPU, 16 GB of running memory, and the GPU is NVIDIA RTX3060TI. The deep learning framework pytorch is used to recognize medicinal herbs. This article conducts experimental design on the proposed model. Firstly, based on the actual situation of traditional Chinese medicine recognition, the accuracy of classification is selected to measure whether the experimental results meet the requirements. Accuracy refers to the percentage of correctly predicted quantities in the entire dataset. The analysis of the experimental results is the recognition accuracy of the proposed model on the dataset.

This article is an improvement on the ConvNeXt network architecture by adding Stacked ACMix to improve network performance. At the same time, it improves the ability to extract information from features. Due to the inclusion of CNN and Attention modules in ACMix, it is possible to extract local information from images as well as global information features through the Multi Head Self Attention module. Secondly, a comparative experiment is conducted on network models of different depths, with models of 20-layer blocks, 22-layer blocks, and 24-layer blocks designed for comparison, and a network model with better classification performance is selected.

4.2 Analysis of experimental results

In order to improve the recognition rate of the model, the Stacked ACMix module is added when building the network model. In order to verify the effect of the Stacked ACMix module, a 20-layer block neural network model is used, and the GELU is the activation function. A comparative experiment is designed to determine whether to add the Stacked ACMix module. Our data set has 7,853 images, including 5,497 images in the training set, 1,571 images in the verification set, and 785 images in the test set. Figure 5 shows the corresponding recognition rates trained after adding the Stacked ACMix module in the network model, verified after adding the Stacked ACMix module, trained without adding the Stacked ACMix module, and verified without adding the Stacked ACMix module layer.

FIGURE 5
www.frontiersin.org

FIGURE 5. Recognition results.

From the analysis of the above experimental results, it can be concluded that adding the Stacked ACMix module to the network model significantly improves the recognition rate. After adding the Stacked ACMix module, the training and verification rates reached 89.3% and 83.1%, respectively, while the training and verification rates without the Stacked ACMix module were 75.3% and 69.6%. Therefore, the Stacked ACMix module will improve the performance of the model and improve its recognition accuracy.

In CNN, insufficient depth of the model will make it impossible to correctly fit the features of the data. If the model is too deep, it will require learning a large number of weight parameters to increase the training time of the model (He et al., 2016). Complex models are prone to overfitting in the training process, reducing the generalization ability of the model, so it is necessary to combine the actual situation to construct the depth of the model.In order to explore the recognition rate between different depth models, this article conducted medicinal image recognition experiments on three different models: 20-layer, 22-layer, and 24-layer. The results in Figure 6 shows the training and verify rates of the 20-layer, 22-layer, and 24-layer networks. From the analysis of the following experimental results, it can be concluded that when the depth of the network model is 22-layers, the recognition rate of the model is optimal, with a training rate of 91.3% and a verify rate of 85.2%, respectively. Secondly, when the depth of the model is 20 layers, the training rate and verify rate are 81.6% and 75.3%, respectively. Finally, when the depth of the model is 24 layers, the classification performance of the model is poor, with training and verify rates of 79.6% and 71.1%, respectively. Therefore, when the depth of the model is 22-layers, the model has good classification performance. Supplementary Table S1 shows the result of a model based on a 22-layer convolutional neural network on a test set is 80.5%. From the above analysis, it can be concluded that when the model depth is 22-layers, the model has a higher recognition rate on the training set, validation set, and test set. Therefore, we use a model depth of 22-layers as the number of layers for the convolutional neural network model.

FIGURE 6
www.frontiersin.org

FIGURE 6. Training results.

5 Conclusion

With the development of the times, traditional Chinese medicine identification methods have been continuously inherited and innovated, and deep learning technology has been applied to the identification process of traditional Chinese medicine to obtain more objective, simple and convenient practical identification methods. This article first introduces the categories of traditional Chinese medicine, and then preprocesses the medicinal image through normalization, grayscale, noise reduction, and other methods to improve the effectiveness of the medicinal image recognition algorithm. Data enhancement technology is used to increase the data size of the medicinal image and solve the problem of a small image dataset. Finally, this article uses the improved ConvNeXt to extract features and classify and differentiate traditional Chinese medicine. By adding ACMix and stacked FFN networks to the ConvNeXt network, its feature extraction ability is improved. The efficiency and accuracy of the model are improved by setting different depth network models for comparison. The specific recognition effect of traditional Chinese medicine images is tested using the constructed neural network model.

Data availability statement

The raw data supporting the conclusion of this article will be made available by the authors, without undue reservation.

Author contributions

JM: study conception and administration. JM, YH, and ZhW: methodology and validation. YH, ZeW, and JL: experimental work and manuscript drafting. JM and JL: manuscript review and editing. All authors contributed to the article and approved the submitted version.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fbioe.2023.1199803/full#supplementary-material

References

Altuntaş, Y., Co mert, Z., and Kocamaz, A. F. (2019). Identification of haploid and diploid maize seeds using convolutional neural networks and a transfer learning approach. Comput. Electron. Agric. 163, 104874. doi:10.1016/j.compag.2019.104874

CrossRef Full Text | Google Scholar

Artzai, P., Maximiliam, S., and Aitor, A. G. (2019). Crop conditional Convolutional Neural Networks for massive multi-crop plant disease classification over cell phone acquired images taken on real field conditions[J]. Comput. Electron. Agric. 2019.

Google Scholar

Batchuluun, G., Naqvi, R., Kim, W., and Park, K. (2018). Body-movement-based human identification using convolutional neural network. Expert Syst. Appl. 101, 56–77. doi:10.1016/j.eswa.2018.02.016

CrossRef Full Text | Google Scholar

Chen, H., and He, Y. (2022). Machine learning approaches in traditional Chinese medicine: A systematic review. Am. J. Chin. Med. 50 (01), 91–131. doi:10.1142/s0192415x22500045

PubMed Abstract | CrossRef Full Text | Google Scholar

Chen, W., Tong, J., He, R., Lin, Y., Chen, P., Chen, Z., et al. (2021). An easy method for identifying 315 categories of commonly-used Chinese herbal medicines based on automated image recognition using AutoML platforms. Inf. Med. Unlocked 25, 100607. doi:10.1016/j.imu.2021.100607

CrossRef Full Text | Google Scholar

Chu, Z., Li, F., Wang, D., Shusheng, X., Chunfeng, G., and Haoran, B. (2022). Research on identification method of tangerine peel year based on deep learning[J]. Food Sci. Technol. 2022, 42. doi:10.1590/fst.64722

CrossRef Full Text | Google Scholar

Dosovitskiy, A., Beyer, L., and Kolesnikov, A. (2020). An image is worth 16x16 words: Transformers for image recognition at scale[J].

Google Scholar

Fricker, G. A., Ventura, J. D., Wolf, J. A., North, M. P., Davis, F. W., and Franklin, J. (2019). A convolutional neural network classifier identifies tree species in mixed-conifer forest from hyperspectral imagery. Remote Sens. 11 (19), 2326. doi:10.3390/rs11192326

CrossRef Full Text | Google Scholar

Fukushima, K., and Miyake, S. (1982). “Neocognitron: A self-organizing neural network model for a mech-anism of visual pattern recognition[C],” in Proceedings of Competition and cooperation in neural nets, Kyoto, Japan, February 15–19, 1982, 267–285.

CrossRef Full Text | Google Scholar

Han, M., Zhang, J., Zeng, Y., Hao, F., and Ren, Y. (2022). A novel method of Chinese herbal medicine classification based on mutual learning. Mathematics 10 (9), 1557. doi:10.3390/math10091557

CrossRef Full Text | Google Scholar

Hansena, M., Smitha, M., Smitha, L., Salter, M. G., Baxter, E. M., Farish, M., et al. (2018). Towards on-farm pig face recognition using convolutional neural networks. Comput. Industry 98, 145–152. doi:10.1016/j.compind.2018.02.016

CrossRef Full Text | Google Scholar

Hao, W., Han, M., and Li, S. (2022). Mtal: A novel Chinese herbal medicine classification approach with mutual triplet attention learning[J]. Wirel. Commun. Mob. Comput. 2022.

CrossRef Full Text | Google Scholar

Hao, W., Han, M., Yang, H., Hao, F., and Li, F. (2021). A novel Chinese herbal medicine classification approach based on EfficientNet. Syst. Sci. Control Eng. 9 (1), 304–313. doi:10.1080/21642583.2021.1901159

CrossRef Full Text | Google Scholar

He, K., Zhang, X., and Ren, S. (2016). “Deep residual learning for image recognition[C],” in Proceedings of the IEEE conference on computer vision and pattern recognition, Las Vegas, Nevada, 27-30 June 2016, 770–778.

Google Scholar

He, K., Zhang, X., Ren, S., and Sun, J. (2016). “Deep residual learning for image recognition[C],” in Proceedings of the IEEE conference on computer vision and pattern recognition, Las Vegas, NV, USA, 27-30 June 2016, 770–778.

Google Scholar

Hu, J., Chen, Z., Yang, M., Zhang, R., and Cui, Y. (2018). A multiscale fusion convolutional neural network for plant leaf recognition. IEEE Signal Process. Lett. 25 (6), 853–857. doi:10.1109/lsp.2018.2809688

CrossRef Full Text | Google Scholar

Huang, M. L., Xu, Y. X., and Liao, Y. C. (2021). Image dataset on the Chinese medicinal blossoms for classification through convolutional neural network. Data Brief 39, 107655. doi:10.1016/j.dib.2021.107655

PubMed Abstract | CrossRef Full Text | Google Scholar

Hubel, D. H., and Wiesel, T. N. (1962). Receptive fields, binocular interaction and functional architecture in the cat's visual cortex. J. physiology 160 (1), 106–154. doi:10.1113/jphysiol.1962.sp006837

CrossRef Full Text | Google Scholar

Hussein, B. R., Malik, O. A., Ong, W. H., and Slik, J. W. F. (2020). Automated classification of tropical plant species data based on machine learning techniques and leaf trait measurements. Comput. Sci. Technology/6th ICCST 2019 Lect. Notes Electr. Eng. 2020, 85–94.

CrossRef Full Text | Google Scholar

Jeong, D., Kim, B. G., and Dong, S. Y. (2020). Deep joint spatiotemporal network (DJSTN) for efficient facial expression recognition. Sensors 20 (7), 1936. doi:10.3390/s20071936

PubMed Abstract | CrossRef Full Text | Google Scholar

Kattenborn, T., Eichel, J., and Fassnacht, F. E. (2019). Convolutional Neural Networks enable efficient, accurate and fine-grained segmentation of plant species and communities from high-resolution UAV imagery. Sci. Rep. 9, 17656. doi:10.1038/s41598-019-53797-9

PubMed Abstract | CrossRef Full Text | Google Scholar

Kim, J. H., Kim, B. G., Roy, P. P., and Jeong, D. M. (2019). Efficient facial expression recognition algorithm based on hierarchical deep neural network structure. IEEE access 7, 41273–41285. doi:10.1109/access.2019.2907327

CrossRef Full Text | Google Scholar

Krizhevsky, A., Sutskever, I., and Hinton, G. E. (2012). Imagenet classification with deep convolutional neural networks[J]. Adv. neural Inf. Process. Syst. 25, 1097–1105.

Google Scholar

LeCun, Y., Bottou, L., Bengio, Y., and Haffner, P. (1998). Gradient-based learning applied to document recognition. Proc. IEEE 86, 2278–2324. doi:10.1109/5.726791

CrossRef Full Text | Google Scholar

Lee, S. H., Chan, C. S., Wilkin, P., and Remagnino, P. (2015). “Deep-plant: Plant identification with convolutional neural networks[C],” in 2015 IEEE International Conference on Image Processing (ICIP), Quebec City, QC, Canada, 27-30 September 2015, 452–456.

Google Scholar

Liu, P., Lei, T., Xiang, Q., Wang, Z., and Wang, J. (2022). Animal Fur recognition algorithm based on feature fusion network. J. Multimedia Inf. Syst. 9 (1), 1–10. doi:10.33851/jmis.2022.9.1.1

CrossRef Full Text | Google Scholar

Namatēvs, I. (2017). Deep convolutional neural networks: Structure, feature E×traction and training[J]. Nephron Clin. Pract. 20 (1), 40–47.

Google Scholar

Park, S-J., Kim, B-G., and Chilamkurti, N. (2021). A robust facial expression recognition algorithm based on multi-rate feature fusion scheme. Sensors 21 (21), 6954. doi:10.3390/s21216954

PubMed Abstract | CrossRef Full Text | Google Scholar

Park, Y., Kim, H., Kim, D., Lee, H., Kim, S. B., and Kang, P. (2017). A deep learning-based sports player evaluation model based on game statistics and news articles. Knowledge-Based Syst. 138, 15–26. doi:10.1016/j.knosys.2017.09.028

CrossRef Full Text | Google Scholar

Shotton, J., Girshick, R., Fitzgibbon, A., Sharp, T., Cook, M., Finocchio, M., et al. (2013). Efficient human pose estimation from single depth images. IEEE Trans. Pattern Analysis Mach. Intell. 35 (12), 2821–2840. doi:10.1109/tpami.2012.241

CrossRef Full Text | Google Scholar

Simonyan, K., and Zisserman, A. (2015). Very deep convolutional networks for large-scale image recognition[C]. San Diego: IEEE, 1–14. International Conference on Learning Representations.

Google Scholar

Simonyan, K., and Zisserman, A. (2014). Very deep convolutional networks for large-scale image recognition[J]. ar Xiv.

Google Scholar

Sun, X., and Qian, H. (2016). Chinese herbal medicine image recognition and retrieval by convolutional neural network[J]. Plos one 11 (6), 1–19.

CrossRef Full Text | Google Scholar

Szegedy, C., Liu, W., Jia, Y., Sermant, P., Reed, S., Anguleov, D., et al. (2015). “Going deeper with convolutions[C],” in Proceedings of the IEEE conference on computer vision and pattern recognition, Boston, MA, USA, 07-12 June 2015, 1–9.

Google Scholar

Wang, W., Tian, W., Liao, W., Cai, B., and Li, B. (2021). “Identifying Chinese herbal medicine by image with three deep CNNs[C],” in Proceedings of the 5th International Conference on Control Engineering and Artificial Intelligence, 1-8 January 2021.

Google Scholar

Wang, Y., Hao, C., Li, Y., Chen, P., Zhang, P., Wu, H., et al. (2020). NF-κB/TWIST1 mediates migration and phagocytosis of macrophages in the mice model of implant-associated Staphylococcus aureus osteomyelitis. J. Comput. Appl. 40 (5), 1301. doi:10.3389/fmicb.2020.01301

CrossRef Full Text | Google Scholar

Wang, Y., Tao, S., Li, Y., Wang, W., Wang, S., Duan, J., et al. (2023). Clinical impact and risk factors of intensive care unit-acquired nosocomial infection: A propensity score-matching study from 2018 to 2020 in a teaching hospital in China. SPIE 12462, 569–579. doi:10.2147/IDR.S394269

CrossRef Full Text | Google Scholar

Wu, L., Xing, Y., and Zheng, B. (2016). A preliminary study of zanthoxylurn bungeanum maxim varieties discriminating by computer vision[J]. Chin. J. Sensors Actuators 29 (1), 136–140.

Google Scholar

Xing, C., Huo, Y., Huang, X., Lu, C., Liang, Y., Wang, A., et al. (2020). “Research on image recognition technology of traditional Chinese medicine based on deep transfer learning[C],” in 2020 International Conference on Artificial Intelligence and Electromechanical Automation (AIEA), Tianjin, China, June 26 2020 to June 28 2020, 140–146.

Google Scholar

Xu, Y., Wen, G., Hu, Y., Luo, M., Dai, D., Zhuang, Y., et al. (2021). Multiple attentional pyramid networks for Chinese herbal recognition. Pattern Recognit. 110, 107558. doi:10.1016/j.patcog.2020.107558

CrossRef Full Text | Google Scholar

Liu, Z., Mao, H., Wu, C-Y., Feichtenhofer, C., Darrell, T., and Xie, S. (2022). A ConvNet for the 2020s. arXiv 2201.03545.

CrossRef Full Text | Google Scholar

Wang, L., Hu, K., and Zhang, Y. (2019). Factor graph model based user profile matching across social networks[J]. IEEE Access 7, 152429–152442. doi:10.1109/ACCESS.2019.2948073

CrossRef Full Text | Google Scholar

Wang, L., Zhang, K., and Hu, K. (2021). FEUI: Fusion Embedding for User Identification across social networks. Neural Computing and Applications 52, 8209–8225. doi:10.1007/s10489-021-02716-5

CrossRef Full Text | Google Scholar

Wang, L., Zhang, Y., Yuan, J., Hu, K., and Cao, S. (2022). FEBDNN: Fusion Embedding Based Deep Neural Network for user retweeting behavior prediction on social networks. Neural Computing and Applications 34 (16), 13219–13235. doi:10.1007/s00521-022-07174-9

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang, S. W., Zhang, S. B., Zhang, C. L., Wang, X., and Shi, Y. (2019). Cucumber leaf disease identification with global pooling dilated convolutional neural network. Comput. Electron. Agric. 162, 422–430. doi:10.1016/j.compag.2019.03.012

CrossRef Full Text | Google Scholar

Zhao, J. H., Chen, X. H., and Dou, X. T. (2022). Automatic classification of medicinal materials based on three-dimensional point cloud and surface spectral information[C]//International Conference on Computer Graphics, Artificial Intelligence, and Data Processing (ICCAID 2021). SPIE 12168, 384–394.

Google Scholar

Zhou, W., Yang, K., Zeng, J., Lai, X., Wang, X., Ji, C., et al. (2021). FordNet: Recommending traditional Chinese medicine formula via deep neural network integrating phenotype and molecule. Pharmacol. Res. 173, 105752. doi:10.1016/j.phrs.2021.105752

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhu, X., Zhu, M., and Ren, H. (2018). Method of plant leaf recognition based on improved deep convolutional neural network. Cognitive Syst. Res. 52, 223–233. doi:10.1016/j.cogsys.2018.06.008

CrossRef Full Text | Google Scholar

Keywords: traditional Chinese medicinal, deep learning, classification, image recognition, network models

Citation: Miao J, Huang Y, Wang Z, Wu Z and Lv J (2023) Image recognition of traditional Chinese medicine based on deep learning. Front. Bioeng. Biotechnol. 11:1199803. doi: 10.3389/fbioe.2023.1199803

Received: 04 April 2023; Accepted: 17 May 2023;
Published: 21 July 2023.

Edited by:

B. D. Parameshachari, Nitte Meenakshi Institute of Technology, India

Reviewed by:

Suyel Namasudra, International University of La Rioja, Spain
Jun Hu, Rey Juan Carlos University, Spain
Byung-Gyu Kim, Sookmyung Women’s University, Republic of Korea

Copyright © 2023 Miao, Huang, Wang, Wu and Lv. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Yanan Huang, hynemma@126.com

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.