- 1Agricultural Economics and Information Institute, Jiangxi Academy of Agriculture Sciences, Nanchang, China
- 2AgroBioChem, Precision Livestock and Nutrition Unit, Gembloux Agro-Bio Tech, University of Liège, Gembloux, Belgium
- 3Information Technology Group, Wageningen University and Research, Wageningen, Netherlands
- 4Agricultural Information Institute, Chinese Academy of Agriculture Sciences, Beijing, China
Automatic monitoring of feeding behavior especially rumination and eating in cattle is important to keep track of animal health and growth condition and disease warnings. The noseband pressure sensor is not only able to accurately sense the pressure change of the cattle's jaw movements, which can directly reflect the cattle's chewing behavior, but also has strong resistance to interference. However, it is difficult to keep the same initial pressure while wearing the pressure sensor, and this will pose a challenge to process the feeding behavior data. This article proposed a machine learning approach aiming at eliminating the influence of initial pressure on the identification of rumination and eating behaviors. The method mainly used the local slope to obtain the local data variation and combined Fast Fourier Transform (FFT) to extract the frequency-domain features. Extreme Gradient Boosting Algorithm (XGB) was performed to classify the features of rumination and eating behaviors. Experimental results showed that the local slope in combination with frequency-domain features achieved an F1 score of 0.96, and recognition accuracy of 0.966 in both rumination and eating behaviors. Combined with the commonly used data processing algorithms and time-domain feature extraction method, the proposed approach improved the behavior recognition accuracy. This work will contribute to the standardized application and promotion of the noseband pressure sensors.
Introduction
Precision livestock farming (PLF) is a research field involving multiple disciplines such as the Internet of Things (IoT) and artificial intelligence (AI). Through the continuous real-time monitoring of individual livestock's health and growth condition, not only the animal welfare but also the production and quality can be improved further (1, 2). Feeding behavior is one of the key indicators in cattle to measure growth and diseases. Accurate analysis of feeding behavior can help evaluate the feed intake and growth rate, which could be used to provide a reference for cattle breeding (3). In particular, rumination and eating are the most direct and effective feeding behavior characteristics to confirm the cattle's health status (4). Therefore, automatic and accurate monitoring of rumination and eating in cattle is of great significance to building precision livestock.
In traditional livestock farming, direct observation is one of the most widely used methods (5). Stopwatch, counter, telescope, and other tools are commonly used to track cattle for continuous recording of their eating, rumination, resting, and wandering behaviors. Unfortunately, these methods are both time and labor-consuming, especially for the grazing sector (6). With the development of information technologies, sensor technology has been applied to monitor cattle behaviors to achieve automatic and continuous detection, which is more efficient and less intrusive than the traditional manual monitoring method (7). Due to the favorable stability and endurance, wireless sensors can provide long-term monitoring, and therefore, are promising in grazing pastures (8). The sensors consist of sound sensors (9, 10), acceleration sensors (11, 12), and pressure sensors (13, 14). The sound sensors are mainly used to detect the chewing behaviors of cattle such as chewing, biting, and chew-bite through the sound produced during chewing (15, 16). For example, Chelotti et al. (17) proposed an online bottom-up foraging activity recognizer algorithm incorporating multilayer perceptron (MLP) along with a decision tree and achieved the F1 scores of 82.2% (grazing) and 74.3% (rumination) in the 5 min detection window size. Although the sound sensors have good performance for monitoring chewing behavior in the ideal environment, they are susceptible to being affected by noise in complex farms (18). The monitoring method based on acceleration sensors is to fix the devices on the head, mandible, ear, neck, or other parts of cattle and then identify the feeding behaviors by distinguishing the movements and postures of the acceleration (19, 20). For instance, Smith et al. (7) used the triaxial acceleration to collect cattle motion data signals and the one-VS-all classification framework was proposed to recognize grazing, walking, ruminating, resting, and other behaviors. In the 30 s window size condition, this method achieved the F1 score of 0.98 (grazing) and 0.86 (ruminating). The acceleration sensors are always affected by the semblable acceleration signals and different behaviors may have semblable signal features, it is difficult to judge the behaviors for acceleration-based models, especially in practical scenarios (21). Given that sound sensors and acceleration sensors have limitations in the task of behavior recognition, some studies have improved the classification performance by combining multiple sensors (22, 23).
The environment of pasture is more complex than that inside farmhouses, and thus, it is more challenging to monitor the feeding behavior. Compared with sound and acceleration sensors, noseband pressure sensors can directly sense the pressure change of cattle's jaw movements and reflect the chewing behavior, which is a direct and effective way to monitor real-time feeding behavior (24). Since the pressure signal of sensors and cattle's chewing behavior have a very high correlation (25). Rutter (26) proposed the program (called Graze) that was used to apply the amplitude and frequency of pressure data to conduct the behavior identification of eating and rumination and obtained 91% identification accuracy. However, Graze was used in conjunction with the IGER Behavior Recorder, and the recorder interfered with the animals. Zehner et al. (27) developed and validated a novel scientific monitoring device for automated measurement of ruminating and eating behavior in stable-fed cows to provide research with a measuring instrument, and published two software versions of RumiWatch to identify cattle foraging behavior, and the average recognition accuracy of the two software versions was 88.86 and 85.31% under the condition of 1 min windows. Subsequently, several versions of RumiWatch were released, and the prediction accuracy of ruminant behavior and foraging behavior in the 1 min window were both higher than 90% (28, 29). Shen et al. (30) applied the pressure sensor to count rumination bouts, duration of rumination, and the number of cuds, respectively, and obtained an accuracy of 100, 94.2, and 94.45%. They proposed to use the SD and spectrum characteristics of pressure data for rumination behavior recognition and achieved an accuracy of 94.2% under the condition of 51.2 s time resolution.
The monitoring of feeding behavior could help predict herbage demand (31). Consequently, to monitor the cattle's health more effectively and improve animal welfare, the feeding behavior recognition algorithm based on pressure sensors needs more progress for future potential applications (32). However, in previous studies, the pressure sensor-based feeding behavior recognition algorithm is built on the peak rates and the peak intervals of cattle chewing data. The peak is detected by the threshold, which is very sensitive and easily affected by the pressure values. Due to the difference in the size of cattle head, to obtain the initial state (trough of the pressure value) of the chewing process, it is necessary to ensure that the jaw is fully occluding, and the flexible band is at the same stretch degree. To achieve this, there is a need to observe the jaw movement and the pressure value of the equipment at the same time and then adjust the equipment accordingly, which is quite laborious. Furthermore, the maximum pressure of complete opening is still out of control. Moreover, the classification of feeding behavior is based on general algorithms without animal-specific learning data, and thus the average recognition accuracy is low.
Machine learning (ML) methods have been widely considered in feeding behavior recognition, namely, support vector machine (SVM), random forest (RF), and extreme gradient boosting (XGBoost) (33). Dutta et al. (34) adopted supervised machine learning techniques for cattle behavioral classification with a 3-axis accelerometer and magnetometer, and the highest average correct classification accuracy of 96% was achieved using the bagging ensemble classification with Tree learner. Fogarty et al. (35) explored four ML algorithms (CART, SVM, LDA, and QDA) for sheep behavior classification with ear-borne accelerometers, and the accuracy for each ethogram was over 75%. Riaboff et al. (36) developed a prediction method for feeding behavior and posture using accelerometer data based on the XGB algorithm and also presented a variety of machine learning algorithms for comparisons. Balasso et al. (3) developed a model to identify posture and behavior from the data collected from a triaxial accelerometer located on the left flank of dairy cows, and four algorithms (RF, KNN, SVM, and XGB) were tested and the XGB model showed the best accuracy. Dutta et al. (37) developed and deployed a neck-mounted intelligent IoT device for cattle monitoring using the XGBoost classifier, which achieved an overall classification accuracy of 97%. Although XGB algorithms achieved good performance in accelerometer data analysis, the XGB algorithms with noseband pressure sensors have never been applied to the classification of livestock behavior. Therefore, there is a need to assess the application of the XGB algorithm to recognize livestock behaviors with noseband pressure data.
To simplify the standardizing of the wearing of the nasal band pressure sensor and improve feeding behavior recognition, this article aims to explore a data processing approach in combination with a machine learning model to eliminate the impact of different ranges of noseband pressure values and facilitate the behaviors recognition.
Materials and Methods
Experimental Equipment
This research developed a set of comprehensive data acquisition equipment, which could collect masticatory pressure data and record behavior data in videos at the same time. The equipment consisted of the masticatory pressure collecting device, the miniature camera (Gold touch U21), and the adjustable wearable collar, as shown in Figure 1. The Masticatory pressure collection device included a pressure sensor (0–2,000.000 g), an HX711 ADC converter (sampling frequency 50 Hz), an SoC chip ASR6501 integrated with LoRa, a 16 GB data recorder module, a 3D printing shell, and two lithium batteries (3.7 V, 500 mAh) in Figure 2. Three stretchable cords attached the device to the head of a cow. Since it is very difficult to track the chewing behavior of cattle manually under natural grazing conditions, a micro camera (64 GB SD card), and mobile power supply (5 V, 10,000 mAh) were used to construct a wearable video monitoring system for chewing behavior by 3D printer, to ensure that the whole process of cattle chewing behavior was collected for a long time. The video monitoring system can maintain 24 h of uninterrupted power supply, in which the working currents of data acquisition equipment and miniature cameras are 26 and 230 mA. The mobile power not only supplies power to the miniature camera but also charges the masticatory pressure collecting device. Changing the mobile power supply once a day can maintain the continuous collection of pressure data and video data without dismantling equipment.
Figure 1. Components of the wearable integrated data acquisition device. (a) Masticatory pressure collecting device, (b) miniature camera, and (c) adjustable wearable collar.
Data Collection and Preparation
This research was conducted in the Gao'an base of the Jiangxi Academy of Agricultural Sciences. Three Simmental × Chinese Yellow crossbred cattle, each weighing about 650 kg on average, were used for grazing in the experimental area. Before binding to the cattle, the noseband pressure sensor device was turned on and automatically calibrated for an initial pressure value of 0 g for all devices. At the same time, the breeders helped make the cattle gentle, so that the noseband pressure sensor device and micro-camera could be installed. In addition, the equipment adaptability test on all three cows lasted for 2 days. The formal experiment period lasted for 5 days, from 8 a.m. to 5 p.m. every day. A total of 135 h of masticatory pressure data and the corresponding video monitoring data were collected. Based on the monitoring data of masticatory pressure collected by the micro-camera, the data of masticatory pressure were manually marked. As shown in Table 1, according to the suggestions from animal husbandry experts, the marking rules of cattle feeding behavior were summarized. Then, the annotated pressure data was saved in CSV files, where each line included the recording time, pressure value, and the label. It is worth noting that the noseband pressure sensor device was set up to synchronize with the micro-camera, ensuring that the behavior of the video annotation was corresponding to the original pressure data.
Data Preprocessing
The scope of original pressure data of different individuals varies greatly, and such scale difference has a great influence on the data modeling (38). In this section, data preprocessing was designed to obliterate the scale difference which was produced by the original pressure data. In the mastication data collected from cattle, different initial pressure values were generated after wearing the noseband pressure sensor due to the size difference of the cattle head. This initial pressure value is a relatively stable constant throughout the wearing process of the device. There are two ways to eliminate this constant, one is to extract local changes in the data, and the other is signal filtering. In this study, first-order difference and local slope were used to extract local variation of data. At the same time, the high-pass filter was used to filter out unstable initial variables. Furthermore, the control method exerted in this article means the original pressure data without any processing.
First-Order Difference
First-order difference (FOD) is the simplest way to extract the continuous numerical changes for a given sequence. More specifically, each term in the original sequence is subtracted from the next. After the above operation, the number of arrays is reduced by 1. Specific operations are as follows:
Local Slope
Local slope (LS) is computed using linear regression over the spectral amplitude values (39). It can obtain the variation trend of local signals in samples. The algorithm can eradicate the initial pressure, and this is because the slope of the local signal is solely determined by the trend of signal change. The process of the local slope extraction is to extract all slope features of the data sample according to the sliding window with 11 sampling points and one stride. First, calculating the local slope requires an auxiliary arithmetic array, as shown below:
To make the local slope in agreement with the current window, the least square solution is applied:
is the sample point in each sliding window. The sliding window exhibits the local slope sample of the whole data sample. Figure 3 shows the process diagram of the partial slope extraction preprocessing method in the flow chart. After this process, 20 local slopes are extracted from 0.1 to 0.48 s corresponding to the raw pressure data.
Figure 3. Illustration of the preprocessing operation of partial slope extraction. (A) Aims to extract the slope of the position of 0.1 s and the corresponding least-squares fitting line. (B) Intends to extract the slope of the position of 0.12 s and the corresponding fitting line. (C) It is the local slope after data preprocessing, each point of the local slope corresponding to the original pressure data of local trends. A total of 20 local slopes are extracted excluding the auxiliary calculation points.
High-Pass Butterworth Filter
High-pass Butterworth filter (HBF) is used to remove the frequency region in the waveform whose signal frequency is lower than the cutoff frequency, including the uncertain initial pressure value (40). Specifically, the original signal is processed with a high-pass Butterworth filter with a sixth-order cut-off frequency of 0.3 Hz. The output data points are the same as the input.
Feature Extraction
Features were extracted from two aspects of the time and frequency domain. The frequency-domain feature (FDF) is established on the fast Fourier transform principle (41), as the range of unilateral frequency domain (not normalized). A sequence array of length 250 was extracted as an FDF, exhibiting both the range of intensity in each sampling frequency (0–25 Hz, resolution: 0.1 Hz) and serving as a distributed representation of the preprocessed data. The time-domain statistical feature (TDSF) included position parameters like average, variance, SD, maximum, minimum, range, median, first quartile, third quartile, interquartile range, root mean square, movement variation, skewness, and kurtosis. Skewness and kurtosis extracted in the time domain (34) were computed and the maximum of the pairwise correlations between each axis was considered. The time-domain feature extraction referenced the Riaboff method in the feature extraction process of acceleration sensor data for cattle behavior classification (42). Since the main frequency signal in the process of eating (around 1.3 Hz), ruminating (around 0.7 Hz) has a significant difference in the frequency domain, and time features were widely used in cattle behavior classification. Therefore, this article used FDF and TDSF for the data preprocessing.
Model Training
In this section, the predictions of cow behaviors were carried out using a single classifier to compare the results. Extreme Boosting Algorithm (XGB) is a supervised ensemble learning algorithm (33). Its principle is to output the prediction probability of the corresponding category based on the training samples with multiple features and to find the category that is most probable to predict the kind of cow feeding behavior. XGB serves as an efficient way to mine latent patterns in data. This framework contains multiple hyperparameters, commonly used as follows:
• n_estimators: number of gradient boosted trees.
• max_depth: maximum tree depth for base learners.
• learning_rate: boosting learning rate.
• gamma: minimum loss reduction required to make a further partition
• on a leaf node of the tree.
• subsample: subsample ratio of the training instance.
• min_child_weight: The minimum sum of the required instance weights in a child.
The complexity of the model is directly influenced by the number of estimators and the max depth of the tree, which means that the greater the number of the trees are, the deeper the trees' depth is, and the more complex the model is. To compare the model capacity in this article, the same hyperparameters were used in all of the classifiers as seen in Table 2.
Model Validation
The performance of eight combined models was tested under the same parameters to find the optimal algorithm for data preprocessing and feature extraction. One way to assess model performance objectively is to utilize the cross-validation technique (43). It is used to prevent overfitting in prediction models, and protect against overfitting in a predictive model, particularly in a case where the amount of data may be limited. The major step of cross-validation is to divide all the data into five subsets, each of equal size. Four subsets were used for training, and the left one was used for testing. To ensure that no data were omitted in the prediction, the procedure was repeated 5 times and the confusion matrix was obtained for each time, which was then computed to form a total confusion eventually (44).
Accuracy, that is the proportion of successfully classified samples in total samples, serves as the standard for assessing the overall performance of classifiers. On the other hand, accuracy also reflects the ratio of correctly classified samples to total samples. To evaluate the prediction of true positive (TP), false positive (FP), and false negative (FN) in each scenario, precision, recall, and F1 score were also computed. The calculation of evaluation indicators is as follows:
Results
Multi-Window Performance Results
Noseband pressure sensors have been used to classify feeding behaviors, but classification accuracy and computational load are influenced by signal segmentation. Therefore, this article considers the accuracy and execution time as standard evaluation criteria to evaluate variable window size.
Table 3 shows the results of 1,000 samples with different window sizes (5, 10, 15, 20, and 30 s) and different method combinations. As can be seen, the accuracy achieved higher with the window size increased, but with the longer execution cost. The window sizes of 5 s led to significantly lower accuracy than with other window sizes but the execute time was also the fastest with 11.592 s. Despite the higher accuracy the larger windows size performed, considering the processing speed for potential use, this article selected the window size of 10 s for the data segmentation to balance the accuracy (over 95%) and execution cost. The dataset in this work contains a total of 47,482 samples, namely, 14,321 for rumination, 17,978 for eating, and 15,183 for other behaviors. The entire pressure dataset construction process did not use the data overlapping technique, and each sample was used only once in the original annotation data.
Table 3. The effect of window size on the accuracy (mean and standard deviation) and the executing time.
Data Preprocessing Results
During the experiment, although this work was able to initialize the zero-calibration process when the device was not worn and placed horizontally and freely, it was unable to guarantee the same initial pressure value during the wearing process. The difference in initial pressure values of cows wearing the pressure sensor reflected the overall data on the feeding behavior (eating, ruminating, and others) of the three cows. Figure 4A shows the pressure data of three cows' ruminating behavior during the 180 s. Based on the three data preprocessing methods discussed in this work, the data processed by FOD, LS, and HBF are presented in Figures 4B–D. The minimum values of the original pressure waveforms obtained from the three cows were 28.550, 128.240, and 232.320 g, and the corresponding average waveforms values were 145.549, 235.616, and 385.515 g, respectively, as shown in Table 4. Accordingly, the cows with pressure sensors have different initial pressure values. However, the mean values for all three cattle processed by FOD, LS, and HBF were basically close to 0, which means the data preprocessing methods contribute to eliminating the influence of the initial pressure value. In particular, FOD is a kind of algorithm that uses two sampling points to extract the slope, and it has a small amount of calculation and range of variation and also is sensitive to mechanical noise. By comparison, LS adopts 11 sampling points to extract the slope, which takes into account a large range of changing trends, but costs a large amount of calculation. This method can effectively suppress noise and the pressure data are closer to the changing trend of the original. As shown in Figure 4, the FOD algorithm is sensitive to mechanical noise and produces sudden changes during 120–150 s, but the LS algorithm can eliminate the abrupt effects of sensors and behaviors and thus improve the recognition accuracy.
Figure 4. Graph of data pretreatment algorithm. Sample A, sample B, and sample C come from three different cattle. (A) Raw pressure data. (B) Data processed by the first-order difference method (FOD). (C) The local slope (LS) of raw data. (D) Data processed after the high-pass filter (HPF).
Overall Performance of the Model
The prediction accuracy of machine learning models is shown in Figure 5. The models with frequency-domain features show a higher accuracy in recognition than that with time-domain statistical features. With respect to frequency domain features, three data processing algorithms improve the recognition accuracy of the model, reaching 0.012. In the time-domain feature condition, the local variation extraction algorithm (first-order difference and local slope) has improved by 0.047 compared with the empty operation.
Figure 5. Validation results for each model. “None” indicates the control group in data preprocessing without any data processing. FOD, First-order difference; LS, local slope; HPF, high-pass filter; FDF, frequency-domain features; and TDSF, time-domain statistical features.
Comparisons of Machine Learning Algorithms
Multiple supervised classification models are used in this work to compare the performance of k-nearest neighbor (KNN), support vector machine (SVM), decision tree (DT), and XGBoost models on behavior classification. The performance results of various models are presented in Table 5. The highest performance was obtained by combining LS+FDF and XGBoost (accuracy: 0.966 ± 0.010). The model with the lowest performance is the one that uses HPF+TDSF with SVM (accuracy: 0.689 ± 0.091). In the cross-validation, the frequency-domain models are more accurate and robust than the time-domain models. The best performance across all the data processing streams was obtained by XGBoost with an over 89% accuracy.
Discrimination of Every Behavior
Figure 6 shows the total confusion matrix for each model and the recognition performance of every behavior is shown in Table 6. The recognition results of rumination and eating behavior were basically consistent with the overall performance. The model achieved the F1 score of 0.96 for rumination and eating recognition by using the local slope or first-order difference and feature extraction using the frequency-domain feature. The worst recognition model is using time-domain features which achieved a 0.82 F1 score for rumination behavior and a 0.87 F1 score for eating behavior.
Figure 6. The confusion matrix of each model. FOD, First-order difference; LS, local slope; HPF, high-pass filter; FDF, frequency-domain features; and TDSF, time-domain statistical features. (A) FDF, (B) TDSF, (C) FOD + FDF, (D) FOD + TDSF, (E) LS + FDF, (F) LS + TDSF, (G) HPF + FDF, and (H) HPF + TDSF.
Discussion
The jaw movements of cattle are the most obvious features of feeding behavior. The difference between eating and rumination is mainly reflected in the changes in the frequency, amplitude, and trends of jaw movements. Even though the noseband pressure sensor has the advantages in sensing behaviors and strong resistance to non-feeding behaviors, the differences in cattle individuals and resistance to wearing the pressure sensors while wearing the pressure sensors will result in different initial pressure. At this stage, it is unreliable to use the pressure data to process the feeding behavior. Therefore, this article proposed a machine learning approach combined with data preprocessing to provide an effective way for feeding behavior recognition using the pressure sensors in the grazing pasture. The novelty of this article is that the local slope with the combination of the FFT method was used to eliminate the influence of initial pressure, and frequency-domain features were extracted for feeding behavior recognition.
In this article, three data preprocessing methods, namely, first-order difference, high-pass Butterworth filter, and local slope were used to preprocess the raw data with different initial pressure values. Then the frequency-domain and time-domain features were extracted from the preprocessed data and raw data, respectively, and finally, different combination features were applied to the XGB model for behavior classification. Experimental results indict that the accuracy with data preprocessing is significantly higher than that without any preprocessing. Since the initial pressure of the noseband pressure sensor has an impact on the feeding behavior recognition, the influence of the initial pressure was weakened to different extents using the preprocessing methods. The processing of the first-order variance and local slope was reflected in the changing trend of the jaw pressure in the chewing process. During eating and ruminating, chewing speed was different and it can be observed according to the trend of changes in the jaw pressure (45). A high-pass Butterworth filter was used for filtering the low-frequency signal data (lower than 0.3 Hz) which was not the main frequency signal in the process of eating (around 1.1 Hz) and ruminating (around 0.8 Hz). In general, this work is to facilitate the use of the noseband pressure sensor in practice. That is to say, there is no need to consider keeping the same initial pressure while wearing the device. This work will contribute to the promotion of the high-precision chewing behavior perception method.
As for data feature extraction, since the pressure data of feeding behavior shows a strong regularity, the frequency-domain features were better than the time-domain features in feeding behavior recognition. Rumination consists of a series of uniform chewing behaviors that create a regular waveform, while the pattern of eating is irregular. The distribution of the frequency-domain features could present the regularity and provide the XGB classifier with the optimal features. Results indicated that the local slope combined with frequency-domain feature extraction achieved the best performance among three data preprocessing methods, which yielded the accuracy of 0.966, 0.96, and 0.96 on the whole, rumination and eating, respectively.
Concerning the window size, there is no agreed window size to suggest, as it depends on several parameters such as the number of sequences, processing speed, the location of the sensors, and the specific behaviors to detect (6, 23, 46). This article showed that the larger the time window, the higher the identification accuracy, and the longer the execution time took. In the practical scenarios, the raw data will be collected and sent to the server for processing which will bring a lot of power consumption and traffic costs to the wearable detection equipment. A small amount of result data will reduce the data transmission of a single node and in this way, the LORA can be run stably in the low-speed and long-distance communication mode in the future. Therefore, this article balanced the size of the window size and the amount of model calculation and finally chose 10 s as the window size. Although the accuracy is not the highest, it can provide a reference for the embedded feeding behavior recognition model.
During the experiment, the grazing area was relatively empty and the noseband pressure sensor was not damaged during the use. However, if the cattle fight or attack large objects, it is likely to damage the noseband pressure sensor. Therefore, this article will consider using flexible solar cells for power to reduce the power consumption of the whole equipment, which can reduce the volume of the battery in the noseband pressure equipment. Moreover, the outer shell of the equipment will be made of harder material to withstand greater impact force and meanwhile, the hard shell will be covered with a flexible shell to avoid damage to cattle.
Conclusions
Automatic monitoring of feeding behavior especially rumination and eating in cattle is of importance to keep track of animals' health and growth conditions and disease warnings. In this article, the noseband pressure sensor was used to collect the behavior data of cattle, however, it was difficult to achieve the high-precision accuracy of feeding behavior recognition by using the existing algorithms with raw data. For future potential applications, three data preprocessing methods and two feature extraction algorithms were evaluated in this study. It is concluded that the XGB classification model in combination with the local slope and frequency-domain feature achieved the F1 score of 0.96 and accuracy of 0.966 for feeding behavior recognition. The proposed approach is suitable for processing pressure data with a wide range of variations, which can avoid the adjustment of the pressure sensor while wearing the device. This work will help reduce labor consumption and contribute to the standardized application and promotion of the noseband pressure sensors.
Since the difference in breeds and ages of cattle will potentially influence the model performance, future work in this study will extend to more kinds and ages of cows, to have better scalability and ensure that the proposed method applies to all kinds of farms.
Data Availability Statement
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
Ethics Statement
The animal study was reviewed and approved by Jiangxi Academy of Agricultural Sciences, Institute of Animal Husbandry and Veterinary Laboratory, and Animal Ethics Committee. Written informed consent was obtained from the owners for the participation of their animals in this study.
Author Contributions
GC and BX: conceptualization, methodology, software, formal analysis, writing, original–draft preparation, and visualization. GC, BX, and CL: validation. CL and YG: investigation and data curation. CL, YG, and ZC: resources, supervision, and project administration. GC, CL, and HS: writing–review and editing. All the authors contributed to the article and approved the submitted version.
Funding
This research was funded by the Youth Science Foundation Project of Jiangxi Province (20192ACBL21023), Jiangxi Province Research Collaborative Innovation Special Project for Modern Agriculture (JXXTCXNLTS202106), Jiangxi Academy of Agriculture Sciences Innovation Fund for Doctor Launches Project (20181CBS006), National Natural Science Foundation of China, China (32060776), and Jiangxi Province 03 Special Project and 5G Project (20204ABC03A09).
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher's Note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
1. Stygar AH, Gomez Y, Berteselli GV, Dalla Costa E, Canali E, Niemi JK, et al. A systematic review on commercially available and validated sensor technologies for welfare assessment of dairy cattle. Front Vet Sci. (2021) 8:634338. doi: 10.3389/fvets.2021.634338
2. Nóbrega L, Gonçalves P, Antunes M, Corujo D. Assessing sheep behavior through low-power microcontrollers in smart agriculture scenarios. Comput Electr Agric. (2020) 173:105444. doi: 10.1016/j.compag.2020.105444
3. Balasso P, Marchesini G, Ughelini N, Serva L, Andrighetto I. Machine learning to detect posture and behavior in dairy cows: information from an accelerometer on the animal's left flank. Animals. (2021) 11:972. doi: 10.3390/ani11102972
4. Vanrell SR, Chelotti JO, Galli JR, Utsumi SA, Giovanini LL, Rufiner HL, et al. A regularity-based algorithm for identifying grazing and rumination bouts from acoustic signals in grazing cattle. Comput Electr Agric. (2018) 151:392–402. doi: 10.1016/j.compag.2018.06.021
5. Krause M, Beauchemin KA, Rode LM, Farr BI, Nørgaard P. Fibrolytic enzyme treatment of barley grain and source of forage in high-grain diets fed to growing cattle. J Anim Sci. (1998) 76:2912. doi: 10.2527/1998.76112912x
6. Alvarenga FAP, Borges I, Palkovič L, Rodina J, Oddy VH, Dobos RC. Using a three-axis accelerometer to identify and classify sheep behaviour at pasture. Appl Anim Behav Sci. (2016) 181:91–9. doi: 10.1016/j.applanim.2016.05.026
7. Smith D, Rahman A, Bishop-Hurley GJ, Hills J, Shahriar S, Henry D, et al. Behavior classification of cows fitted with motion collars: decomposing multi-class classification into a set of binary problems. Comput Electron Agricul. (2016) 131:40–50. doi: 10.1016/j.compag.2016.10.006
8. Mansbridge N, Mitsch J, Bollard N, Ellis K, Miguel-Pacheco GG, Dottorini T, et al. Feature selection and comparison of machine learning algorithms in classification of grazing and rumination behaviour in sheep. Sensors. (2018) 18:532. doi: 10.3390/s18103532
9. Deniz NN, Chelotti JO, Galli JR, Planisich AM, Larripa MJ, Leonardo Rufiner H, et al. Embedded system for real-time monitoring of foraging behavior of grazing cattle using acoustic signals. Comput Electr Agricul. (2017) 138:167–74. doi: 10.1016/j.compag.2017.04.024
10. Galli JR, Milone DH, Cangiano CA, Martinez CE, Laca EA, Chelotti JO, et al. Discriminative power of acoustic features for jaw movement classification in cattle and sheep. Bioacoustics. (2020) 29:602–16. doi: 10.1080/09524622.2019.1633959
11. Bishop-Hurley G, Henry D, Smith D, Dutta R, Hills J, Rawnsley R, et al. An investigation of cow feeding behavior using motion sensors. In: IEEE International Instrumentation and Measurement Technology Conference (I2MTC) Proceedings. Montevideo: IEEE (2014). p. 1285–90. doi: 10.1109/I2MTC.2014.6860952
12. González LA, Bishop-Hurley GJ, Handcock RN, Crossman C. Behavioral classification of data from collars containing motion sensors in grazing cattle. Comput Electr Agric. (2015) 110:91–102. doi: 10.1016/j.compag.2014.10.018
13. Rutter SM, Champion RA, Penning PD. An automatic system to record foraging behaviour in free-ranging ruminants. Appl Anim Behav Sci. (1997) 54:185–95. doi: 10.1016/S0168-1591(96)01191-4
14. Rutter SM. Graze: a program to analyze recordings of the jaw movements of ruminants. Behav Res Methods Instru Comput. (2000) 32:86–92. doi: 10.3758/BF03200791
15. Chelotti JO, Vanrell SR, Milone DH, Utsumi SA, Galli JR, Rufiner HL, et al. A real-time algorithm for acoustic monitoring of ingestive behavior of grazing cattle. Comput Electr Agric. (2016) 127:64–75. doi: 10.1016/j.compag.2016.05.015
16. Chelotti JO, Vanrell SR, Galli JR, Giovanini LL, Rufiner HL. A pattern recognition approach for detecting and classifying jaw movements in grazing cattle. Comput Electr Agric. (2018) 145:83–91. doi: 10.1016/j.compag.2017.12.013
17. Chelotti JO, Vanrell SR, Martinez Rau LS, Galli JR, Planisich AM, Utsumi SA, et al. An online method for estimating grazing and rumination bouts using acoustic signals in grazing cattle. Comput Electronics in Agriculture. (2020) 173:105443. doi: 10.1016/j.compag.2020.105443
18. Navon S, Mizrach A, Hetzroni A, Ungar ED. Automatic recognition of jaw movements in free-ranging cattle, goats and sheep, using acoustic monitoring. Biosyst Eng. (2013) 114:474–83. doi: 10.1016/j.biosystemseng.2012.08.005
19. Rayas-Amor AA, Morales-Almaráz E, Licona-Velázquez G, Vieyra-Alberto R, García-Martínez A, Martínez-García CG, et al. Triaxial accelerometers for recording grazing and ruminating time in dairy cows: an alternative to visual observations. J Vet Behav. (2017) 20:102–8. doi: 10.1016/j.jveb.2017.04.003
20. Chapa JM, Maschat K, Iwersen M, Baumgartner J, Drillich M. Accelerometer systems as tools for health and welfare assessment in cattle and pigs - a review. Behav Processes. (2020) 181:104262. doi: 10.1016/j.beproc.2020.104262
21. Giovanetti V, Decandia M, Molle G, Acciaro M, Mameli M, Cabiddu A, et al. Automatic classification system for grazing, ruminating and resting behaviour of dairy sheep using a tri-axial accelerometer. Livest Sci. (2017) 196:42–8. doi: 10.1016/j.livsci.2016.12.011
22. Tani Y, Yokota Y, Yayota M, Ohtani S. Automatic recognition and classification of cattle chewing activity by an acoustic monitoring method with a single-axis acceleration sensor. Comput Electr Agric. (2013) 92:54–65. doi: 10.1016/j.compag.2013.01.001
23. Andriamandroso ALH, Lebeau F, Beckers Y, Froidmont E, Dufrasne I, Heinesch B, et al. Development of an open-source algorithm based on inertial measurement units (IMU) of a smartphone to detect cattle grass intake and ruminating behaviors. Comput Electr Agric. (2017) 139:126–37. doi: 10.1016/j.compag.2017.05.020
24. Kröger I, Humer E, Neubauer V, Kraft N, Ertl P, Zebeli Q. Validation of a noseband sensor system for monitoring ruminating activity in cows under different feeding regimens. Livest Sci. (2016) 193:118–22. doi: 10.1016/j.livsci.2016.10.007
25. Ungar ED, Rutter SM. Classifying cattle jaw movements: comparing IGER Behaviour Recorder and acoustic techniques. Appl Anim Behav Sci. (2006) 98:11–27. doi: 10.1016/j.applanim.2005.08.011
26. Rutter SM. Can precision farming technologies be applied to grazing can precision farming technologies be applied to grazing management. In: Paper Presented at The XXII International Grassland Congress (Revitalising Grasslands to Sustain Our Communities). Sydney, NSW: New South Wales Department of Primary Industry (2013).
27. Zehner N, Umstätter C, Niederhauser JJ, Schick M. System specification and validation of a noseband pressure sensor for measurement of ruminating and eating behavior in stable-fed cows. Comput Electr Agric. (2017) 136:31–41. doi: 10.1016/j.compag.2017.02.021
28. Werner J, Leso L, Umstatter C, Niederhauser J, Kennedy E, Geoghegan A, et al. Evaluation of the RumiWatchSystem for measuring grazing behaviour of cows. J Neurosci Methods. (2018) 300:138–46. doi: 10.1016/j.jneumeth.2017.08.022
29. Steinmetz M, von Soosten D, Hummel J, Meyer U, Danicke S. Validation of the RumiWatch Converter V0.7.4.5 classification accuracy for the automatic monitoring of behavioural characteristics in dairy cows. Arch Anim Nutr. (2020) 74:164–72. doi: 10.1080/1745039X.2020.1721260
30. Shen W, Zhang A, Zhang Y, Wei X, Sun J. Rumination recognition method of dairy cows based on the change of noseband pressure. Inform Proc Agric. (2020) 7:479–90. doi: 10.1016/j.inpa.2020.01.005
31. Shafiullah AZ, Werner J, Kennedy E, Leso L, O'Brien B, Umstatter C. Machine learning based prediction of insufficient herbage allowance with automated feeding behaviour and activity data. Sensors. (2019) 19:479. doi: 10.3390/s19204479
32. Li Z, Cheng L, Cullen B. Validation and use of the rumiwatch noseband sensor for monitoring grazing behaviours of lactating dairy cows. Dairy. (2021) 2:104–11. doi: 10.3390/dairy2010010
33. Chen T, Guestrin C. XGBoost. In: Paper Presented at the Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. New York, NY: Association for Computing Machinery (2016).
34. Dutta R, Smith D, Rawnsley R, Bishop-Hurley G, Hills J, Timms G, et al. Dynamic cattle behavioural classification using supervised ensemble classifiers. Comput Electr Agric. (2015) 111:18–28. doi: 10.1016/j.compag.2014.12.002
35. Fogarty ES, Swain DL, Cronin GM, Moraes LE, Trotter M. Behaviour classification of extensively grazed sheep using machine learning. Comput Electr Agric. (2020) 169:105175. doi: 10.1016/j.compag.2019.105175
36. Riaboff L, Poggi S, Madouasse A, Couvreur S, Aubin S, Bédère N, et al. Development of a methodological framework for a robust prediction of the main behaviours of dairy cows using a combination of machine learning algorithms on accelerometer data. Comput Electr Agric. (2020) 169:105179. doi: 10.1016/j.compag.2019.105179
37. Dutta D, Natta D, Mandal S, Ghosh N. MOOnitor: an IoT based multi-sensory intelligent device for cattle activity monitoring. Sensors Actuat A Physical. (2022) 333. doi: 10.1016/j.sna.2021.113271
38. Singh D, Singh B. Investigating the impact of data normalization on classification performance. Appl Soft Comput. (2020) 97:105524. doi: 10.1016/j.asoc.2019.105524
39. Peeters G, Giordano BL, Susini P, Misdariis N, McAdams S. The Timbre Toolbox: extracting audio descriptors from musical signals. J Acoust Soc Am. (2011) 130:2902–16. doi: 10.1121/1.3642604
40. Bayat A, Pomplun M, Tran DA. A study on human activity recognition using accelerometer data from smartphones. Procedia Comput Sci. (2014) 34:450–7. doi: 10.1016/j.procs.2014.07.009
41. Barandas M, Folgado D, Fernandes L, Santos S, Abreu M, Bota P, et al. TSFEL: time series feature extraction library. SoftwareX. (2020) 11:100456. doi: 10.1016/j.softx.2020.100456
42. Riaboff L, Aubin S, Bédère N, Couvreur S, Madouasse A, Goumand E, et al. Evaluation of pre-processing methods for the prediction of cattle behaviour from accelerometer data. Comput Electr Agric. (2019) 165:104961. doi: 10.1016/j.compag.2019.104961
43. Ojala M, Garriga GC. Permutation tests for studying classifier performance. In: Paper Presented at the 2009 Ninth IEEE International Conference on Data Mining. Washington, DC: IEEE Computer Society (2009).
44. Pober DM, Staudenmayer J, Raphael C, Freedson PS. Development of novel techniques to classify physical activity mode using accelerometers. Med Sci Sports Exerc. (2006) 38:1626–34. doi: 10.1249/01.mss.0000227542.43669.45
45. Braun U, Trösch L, Nydegger F, Hassig M. Evaluation of eating and rumination behaviour in cows using a noseband pressure sensor. BMC Vet Res. (2013) 9:1–8. doi: 10.1186/1746-6148-9-164
Keywords: noseband pressure sensor, machine learning, XGB, behavior classification, feeding behaviors
Citation: Chen G, Li C, Guo Y, Shu H, Cao Z and Xu B (2022) Recognition of Cattle's Feeding Behaviors Using Noseband Pressure Sensor With Machine Learning. Front. Vet. Sci. 9:822621. doi: 10.3389/fvets.2022.822621
Received: 26 November 2021; Accepted: 25 April 2022;
Published: 25 May 2022.
Edited by:
Nicole Kemper, University of Veterinary Medicine Hannover, GermanyReviewed by:
Lena Maria Lidfors, Swedish University of Agricultural Sciences, SwedenTomas Norton, KU Leuven, Belgium
Copyright © 2022 Chen, Li, Guo, Shu, Cao and Xu. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Guipeng Chen, Y2hlbmd1aXBlbmcxOTgzJiN4MDAwNDA7MTYzLmNvbQ==; Beibei Xu, ODIxMDExODEyNzAmI3gwMDA0MDtjYWFzLmNu