Skip to main content

EDITORIAL article

Front. Bird Sci.
Sec. Bird Conservation and Management
Volume 3 - 2024 | doi: 10.3389/fbirs.2024.1518077
This article is part of the Research Topic Computational Bioacoustics and Automated Recognition of Bird Vocalizations: New Tools, Applications and Methods for Bird Monitoring View all 6 articles

"Editorial: [Computational Bioacoustics and Automated Recognition of Bird Vocalizations: New tools, Applications and Methods for Bird Monitoring]"

Provisionally accepted
  • 1 Department of Computer Science, Tel-Hai College, Upper Galilee, Israel, Tel-Hai, Israel
  • 2 Ecology Department, University of Alicante, Alicante, Valencian Community, Spain

The final, formatted version of the article will be published soon.

    In recent decades, technological advancements have significantly transformed wildlife monitoring methods, particularly through the adoption of automated and non-invasive techniques like passive acoustic monitoring (PAM). PAM involves deploying sound recorders in the field to capture audio during specified times, which are followed by sound interpretation to gather information on the presence and behavior of species. A major challenge associated with PAM is the vast amount of audio data generated, necessitating the development of automated processing techniques utilizing machine learning, deep learning, or other advanced audio signal processing methods (Stowell, 2022). Birds have been the primary focus of PAM studies, making our understanding of how to effectively monitor and automatically identify bird species from audio recordings well-established.Indeed, PAM has been shown to be a reliable method for estimating bird species richness and population densities from sound recordings (Darras et al., 2019;Pérez-Granados & Traba, 2021).Additionally, it has been employed to investigate various ecological questions, including the detection of previously unnoticed habitat changes and shifts in biodiversity patterns (Ross et al. 2023).However, despite these advances, there remain many challenges and uncertainties that need to be addressed, such as improving automated species identification accuracy, model evaluation, or pairing PAM with other automated techniques. Through this Research Topic, we present recent advancements in tools, methods, and algorithms for automated bird identification to enhance bird monitoring programs utilizing PAM.One of the most important tasks in computational bioacoustics is tracking and monitoring biodiversity changes, which have been greatly advanced using PAM. Among the factors influencing biodiversity, agriculture plays a major role. Employing PAM to assess the impact of agricultural practices on species occupancy and diversity requires careful planning and consideration of key factors for successful research. The work of Molina-Mora et al. discusses these factors and use PAM for assessing the impact of agricultural management practices on biodiversity, using birds as indicators. Specifically, their research focuses on coffee-growing areas in Costa Rica, comparing the effects of pruning and pesticide use over two years. Utilizing a mobile application for community-driven bioacoustic annotations to identify the vocalizations of selected species, they report that pruning negatively affects some species, while pesticide application reduces vocal activity and presumably the presence of all species studied. Currently, most of PAM projects use automated signal recognition software and machine learning models to detect their species of interest (Kahl et al., 2021;Cole et al., 2022;Pérez-Granados, 2023;Lavner et al., 2024;Sethi et al., 2024). Yet, the application of machine learning models faces significant challenges due to the unique characteristics of bioacoustic data and the complexity of real-world deployment. For instance, PAM recordings often have low SNR, as bird vocalizations are embedded in substantial background noise. There is also a mismatch between training data, typically made with directional microphones, and field-recorded PAM data, which captures multiple species simultaneously, often with background noise. Additional challenges include class imbalance, distribution shifts, and inconsistent annotation quality, all of which hinder machine learning models' ability to generalize across varied conditions.Two approaches are presented in the Research Topic to address these challenges: specialized models trained on specific datasets and tailored to answer particular tasks or specific research questions, and generalizable models capable of adapting to a variety of datasets and tasks.A specialized solution is offered by Haley et al. who track the activity of mountain chickadee (Poecile gambeli) in North America, comparing a custom CNN model (Madhusudhana et al., 2021) to BirdNET (Khal et al., 2021). Their custom model, trained exclusively on PAM recordings, exhibited better performance and greater resilience to noise than BirdNET model, which highlights that specific models trained on specialized (i.e. locally trained) datasets may offer an advantage by being more closely aligned with the audio recording data.In construct, Merriënboer et al. advocate for the construction of bioacoustic generalizable or foundation models ("large scale machine learning models that are trained on broad data that can be adapted to a wide variety of downstream tasks", Bommasani et al., 2021), designed to learn from a specific training set or even a small number of examples , (few shot learning (Nolasco et al. 2023;Ghani et al., 2023)) generalize to new species, environments, and tasks beyond those they were initially trained on. They provide a comprehensive review of the challenges associated with developing such models, as well as the methods and metrics used for their evaluation.Another challenge in species classification for PAM is the common practice of setting a threshold to determine whether a species' vocalization is present in an audio clip. Typically, classification scores above the threshold are counted as detections (Pérez-Granados, 2023). Such selection affects the number of false positives or negatives, which can vary across different datasets or species, leading to suboptimal classifications. To resolve this issue, Navine et al. developed a 'threshold-free' bioacoustics analysis framework, which directly estimates call density-the proportion of detection windows containing the target vocalizations-regardless of classifier score, which may be applied to binary detection classifiers with fixed-size windows. A validation method estimates call density in a dataset, and probability distributions are generated for confidence scores in both positive and negative classes. These distributions are used to predict call densities at the site level, accounting for potential shifts in data distribution. Testing on real-world recordings of Hawaiian birds demonstrates this approach's robustness to variations in call density and classifier performance.In PAM surveys it is difficult to estimate the precise location or the identity of vocalizing individuals, though such information is useful for studying birds' behaviour or community dynamics. Guggenberger et al. propose a solution to this problem by coupling a video camera to an acoustic recorder equipped with a 64 microphone array, that enables to localize dozens of individuals, with high temporal and spatial resolution. To demonstrate its potential, they recorded and localized Arabian babblers (Argya squamiceps) during snake-mobbing events. The tagged birds allowed for precise identification of callers and their vocal timing, enabling the reconstruction of vocal social networks. The analysis revealed a periodic pattern in vocalizations, with age-specific inter-call duration, suggesting that snake-mobbing behavior is a learned activity within the group. This topic highlights the potential of PAM for bird monitoring, including studying shifts in bird communities due to human land use, analyzing birds' social networks, and improving the development and assessment of robust, generalizable machine learning models. A significant challenge in advancing automated bird detection models is the scarcity of annotated PAM datasets that span diverse habitats and geographic regions worldwide. To address this, the bioacoustic research community should collaborate to create a global, standardized, and annotated PAM dataset. By pooling resources and adhering to protocols, researchers can promote the development of new, comparable methods for passive bird surveys through the construction and use of scalable foundation models for bioacoustic monitoring, similar to successful applications in other fields (O'Neill et al., 2023). Such efforts would not only advance species identification and acoustic monitoring but also contribute significantly to global nature conservation.

    Keywords: Machine Laerning, deep learning, Passive acoustic monitoring, bioacoustics, Long-term bird monitoring, Biodiversity, Bird vocalisations, audio signal processing

    Received: 27 Oct 2024; Accepted: 06 Dec 2024.

    Copyright: © 2024 Lavner and Pérez-Granados. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

    * Correspondence:
    Yizhar Lavner, Department of Computer Science, Tel-Hai College, Upper Galilee, Israel, Tel-Hai, Israel
    Cristian PĂ©rez-Granados, Ecology Department, University of Alicante, Alicante, 03690, Valencian Community, Spain

    Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.