Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1007/978-3-031-22419-5_19guideproceedingsArticle/Chapter ViewAbstractPublication PagesConference Proceedingsacm-pubtype
Article

Insights from Deep Learning in Feature Extraction for Non-supervised Multi-species Identification in Soundscapes

Published: 04 January 2023 Publication History

Abstract

Biodiversity monitoring has taken a relevant role in conservation management plans, where several methodologies have been proposed to assess biological information of landscapes. Recently, soundscape studies have allowed biodiversity monitoring by compiling all the acoustic activity present in landscapes in audio recordings. Automatic species detection methods have shown to be a practical tool for biodiversity monitoring, providing insight into the acoustic behavior of species. Generally, the proposed methodologies for species identification have four main stages: signal pre-processing, segmentation, feature extraction, and classification. Most proposals use supervised methods for species identification and only perform for a single taxon. In species identification applications, performance depends on extracting representative species features. We present a feature extraction analysis for multi-species identification in soundscapes using unsupervised learning methods. Linear frequency cepstral coefficients (LFCC), variational autoencoders (VAE), and the KiwiNet architecture, which is a convolutional neural network (CNN) based on VGG19, were evaluated as feature extractors. LFCC is a frequency-based method, while VAE and KiwiNet belong to the deep learning area. In ecoacoustic applications, frequency-based methods are the most widely used. Finally, features were tested by a clustering algorithm that allows species recognition from different taxa. The unsupervised approaches performed multi-species identification between 78%–95%.

References

[1]
Pimm SL et al. Emerging technologies to conserve biodiversity Trends Ecol. Evol. 2015 30 685-696
[2]
Dumyahn SL and Pijanowski BC Soundscape conservation Landsc. Ecol. 2011 26 1327-1344
[3]
Sueur J and Farina A Ecoacoustics: the ecological investigation and interpretation of environmental sound Biosemiotics 2015 8 3 493-502
[4]
Aide, T.M., Hern, A., Campos-cerqueira, M.: Species richness (of insects) drives the use of acoustic space in the tropics. Remote Sens. Ecol. Conserv., 1–12 (2017).
[5]
Ross S-J, Friedman NR, Dudley KL, Yoshimura M, Yoshida T, and Economo EP Listening to ecosystems: data-rich acoustic monitoring through landscape-scale sensor networks Ecol. Res. 2017 33 1 135-147
[6]
Ruff ZJ, Lesmeister DB, Duchac LS, Padmaraju BK, and Sullivan CM Automated identification of avian vocalizations with deep convolutional neural networks Remote Sens. Ecol. Conserv. 2020 6 79-92
[7]
Bedoya C, Isaza C, Daza JM, and López JD Automatic recognition of anuran species based on syllable identification Ecol. Inform. 2014 24 200-209
[8]
LeBien J et al. A pipeline for identification of bird and frog species in tropical soundscape recordings using a convolutional neural network Ecol. Inform. 2020 59
[9]
Ruff ZJ, Lesmeister DB, Appel CL, and Sullivan CM Workflow and convolutional neural network for automated identification of animal sounds Ecol. Indic. 2021 124
[10]
Stowell, D.: Computational bioacoustic scene analysis. In: Computational Analysis of Sound Scenes and Events, pp. 303–333. Springer, Cham (2018).
[11]
Xie J, Colonna JG, and Zhang J Bioacoustic signal denoising: a review Artif. Intell. Rev. 2020 54 5 3575-3597
[12]
Noda, J.J., David Sánchez-Rodríguez, C.M.T.-G.: We are IntechOpen, the world’s leading publisher of Open Access books Built by scientists, for scientists TOP 1%. Intech 32, 137–144 (2018)
[13]
Nirosha Priyadarshani, S.M., Castro, I.: Automated birdsong recognition in complex acoustic environments: a review. Avian Biol. (2018).
[14]
Rowe B, Eichinski P, Zhang J, and Roe P Acoustic auto-encoders for biodiversity assessment Ecol. Inform. 2021 62
[15]
Ntalampiras S and Potamitis I Acoustic detection of unknown bird species and individuals CAAI Trans. Intell. Technol. 2021 6 291-300
[16]
Xie J, Hu K, Guo Y, Zhu Q, and Yu J On loss functions and CNNs for improved bioacoustic signal classification Ecol. Inform. 2021 64
[17]
Bedoya, C.L., Molles, L.E.: Acoustic censusing and individual identification of birds in the wild (2021)
[18]
Stowell D Computational bioacoustics with deep learning: a review and roadmap PeerJ 2022 10
[19]
Xie J, Towsey M, Zhu M, Zhang J, and Roe P An intelligent system for estimating frog community calling activity and species richness Ecol. Indic. 2017 82 13-22
[20]
Mermelstein P Distance measures for speech recognition, psychological and instrumental Pattern Recognit. Artif. Intell. 1976 116 374-388
[21]
Zhou, X., Garcia-Romero, D., Duraiswami, R., Carol Espy-Wilson, S.S.: 2011 IEEE Workshop on Automatic Speech Recognition & Understanding: ASRU 2011: Proceedings, Waikoloa, Hawaii, U.S.A., 11–15 December 2011, p. 564 (2011)
[22]
Dong, C., Xue, T., Wang, C.: The feature representation ability of variational autoencoder. Proceedings - 2018 IEEE Third International Conference on Data Science in Cyberspace, DSC 2018, pp. 680–684 (2018).
[23]
Fukumoto, T.: Anomaly detection using Variational Autoencoder (VAE) (2020). https://github.com/mathworks/Anomaly-detection-using-Variational-Autoencoder-VAE-/releases/tag/1.0.1, GitHub. Accessed 23 Apr 2022
[24]
Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. In: 3rd International Conference on Learning Representations ICLR 2015 - Conference Track Proceedings, pp. 1–14 (2015)
[25]
Lamrini B, Le Lann MV, Benhammou A, and Lakhal EK Detection of functional states by the “LAMDA” classification technique: application to a coagulation process in drinking water treatment Comptes Rendus Phys. 2005 6 1161-1168
[26]
Bedoya C, Waissman Villanova J, and Isaza Narvaez CV Gelbukh A, Espinoza FC, and Galicia-Haro SN Yager–Rybalov triple Π operator as a means of reducing the number of generated clusters in unsupervised anuran vocalization recognition Nature-Inspired Computation and Machine Learning 2014 Cham Springer 382-391

Index Terms

  1. Insights from Deep Learning in Feature Extraction for Non-supervised Multi-species Identification in Soundscapes
          Index terms have been assigned to the content through auto-classification.

          Recommendations

          Comments

          Information & Contributors

          Information

          Published In

          cover image Guide Proceedings
          Advances in Artificial Intelligence – IBERAMIA 2022: 17th Ibero-American Conference on AI, Cartagena de Indias, Colombia, November 23–25, 2022, Proceedings
          Nov 2022
          421 pages
          ISBN:978-3-031-22418-8
          DOI:10.1007/978-3-031-22419-5

          Publisher

          Springer-Verlag

          Berlin, Heidelberg

          Publication History

          Published: 04 January 2023

          Author Tags

          1. Feature extraction
          2. Deep learning
          3. Multi-species identification
          4. Biodiversity monitoring
          5. Soundscape

          Qualifiers

          • Article

          Contributors

          Other Metrics

          Bibliometrics & Citations

          Bibliometrics

          Article Metrics

          • 0
            Total Citations
          • 0
            Total Downloads
          • Downloads (Last 12 months)0
          • Downloads (Last 6 weeks)0
          Reflects downloads up to 13 Jan 2025

          Other Metrics

          Citations

          View Options

          View options

          Media

          Figures

          Other

          Tables

          Share

          Share

          Share this Publication link

          Share on social media