4.7 Article

More than a whistle: Automated detection of marine sound sources with a convolutional neural network

Journal

FRONTIERS IN MARINE SCIENCE
Volume 9, Issue -, Pages -

Publisher

FRONTIERS MEDIA SA
DOI: 10.3389/fmars.2022.879145

Keywords

marine soundscapes; CNN - convolutional neural network; passive acoustic monitoring; efficientNet-B0; sound source detection; marine mammal acoustics; Delphinids

Funding

  1. Natural Environmental Research Council [NE/S007210/1]
  2. EU's INTERREG VA Programme

Ask authors/readers for more resources

The effective analysis of Passive Acoustic Monitoring (PAM) data has the potential to determine variations in ecosystem health and species presence in the marine environment. In this study, a deep learning model using cloud computing is trained to extract multi-class marine sound sources, providing support for the conservation of marine mammals and ecosystem monitoring.
The effective analysis of Passive Acoustic Monitoring (PAM) data has the potential to determine spatial and temporal variations in ecosystem health and species presence if automated detection and classification algorithms are capable of discrimination between marine species and the presence of anthropogenic and environmental noise. Extracting more than a single sound source or call type will enrich our understanding of the interaction between biological, anthropogenic and geophonic soundscape components in the marine environment. Advances in extracting ecologically valuable cues from the marine environment, embedded within the soundscape, are limited by the time required for manual analyses and the accuracy of existing algorithms when applied to large PAM datasets. In this work, a deep learning model is trained for multi-class marine sound source detection using cloud computing to explore its utility for extracting sound sources for use in marine mammal conservation and ecosystem monitoring. A training set is developed comprising existing datasets amalgamated across geographic, temporal and spatial scales, collected across a range of acoustic platforms. Transfer learning is used to fine-tune an open-source state-of-the-art 'small-scale' convolutional neural network (CNN) to detect odontocete tonal and broadband call types and vessel noise (from 0 to 48 kHz). The developed CNN architecture uses a custom image input to exploit the differences in temporal and frequency characteristics between each sound source. Each sound source is identified with high accuracy across various test conditions, including variable signal-to-noise-ratio. We evaluate the effect of ambient noise on detector performance, outlining the importance of understanding the variability of the regional soundscape for which it will be deployed. Our work provides a computationally low-cost, efficient framework for mining big marine acoustic data, for information on temporal scales relevant to the management of marine protected areas and the conservation of vulnerable species.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available