Soundscape classification with convolutional neural networks reveals temporal and geographic patterns in ecoacoustic data

Interest in ecoacoustics has resulted in an influx of acoustic data and novel methodologies to classify and relate landscape sound activity to biodiversity and ecosystem health. However, indicators used to summarize sound and quantify the effects of disturbances on biodiversity can be inconsistent w...

Full description

Bibliographic Details
Main Authors: Colin A. Quinn, Patrick Burns, Gurman Gill, Shrishail Baligar, Rose L. Snyder, Leonardo Salas, Scott J. Goetz, Matthew L. Clark
Format: Article
Language:English
Published: Elsevier 2022-05-01
Series:Ecological Indicators
Subjects:
Online Access:http://www.sciencedirect.com/science/article/pii/S1470160X22003028
_version_ 1818010008569249792
author Colin A. Quinn
Patrick Burns
Gurman Gill
Shrishail Baligar
Rose L. Snyder
Leonardo Salas
Scott J. Goetz
Matthew L. Clark
author_facet Colin A. Quinn
Patrick Burns
Gurman Gill
Shrishail Baligar
Rose L. Snyder
Leonardo Salas
Scott J. Goetz
Matthew L. Clark
author_sort Colin A. Quinn
collection DOAJ
description Interest in ecoacoustics has resulted in an influx of acoustic data and novel methodologies to classify and relate landscape sound activity to biodiversity and ecosystem health. However, indicators used to summarize sound and quantify the effects of disturbances on biodiversity can be inconsistent when applied across ecological gradients. This study used an acoustic dataset of 487,148 min from 746 sites collected over 4 years across Sonoma County, California, USA, by citizen scientists. We built a custom labeled dataset of soundscape components and applied a deep learning framework to test our ability to predict these soundscape components: human noise (Anthropophony), wildlife vocalizations (Biophony), weather phenomena (Geophony), Quiet periods, and microphone Interference. These soundscape components allowed us to balance predicting variation in environmental recordings and relative time to build a custom labeled dataset. We used these data to quantify soundscape patterns across space and time that could be useful for environmental planning, ecosystem conservation and restoration, and biodiversity monitoring. We describe a pre-trained convolutional neural network, fine-tuned with our sound reference data, with classification achieving an overall F0.75-score of 0.88, precision of 0.94, and recall of 0.80 across the five target soundscape components. We deployed the model to predict soundscape components for all acoustic data and assess their hourly patterns. We noted an increase in Biophony in the early morning and evening, coinciding with peak animal community vocalization (e.g., dawn chorus). Anthropophony increased during morning/daylight hours and was lowest in the evenings, coinciding with diurnal patterns in human activity. Further, we examined soundscape patterns related to geographic properties at recording sites. Anthropophony decreased with increasing distance to major roads, while Quiet increased. Biophony and Quiet were comparable to Anthropophony at more urban/developed and agriculture/barren sites, while Biophony and Quiet were significantly higher than Anthropophony at less-developed shrubland, oak woodland, and conifer forest sites. These results demonstrate that acoustic classification of broad soundscape components is possible with small datasets, and classifications can be applied to a large acoustic dataset to gain ecological knowledge.
first_indexed 2024-04-14T05:49:58Z
format Article
id doaj.art-eaeda47cfe52422385952dbf25142ecb
institution Directory Open Access Journal
issn 1470-160X
language English
last_indexed 2024-04-14T05:49:58Z
publishDate 2022-05-01
publisher Elsevier
record_format Article
series Ecological Indicators
spelling doaj.art-eaeda47cfe52422385952dbf25142ecb2022-12-22T02:09:09ZengElsevierEcological Indicators1470-160X2022-05-01138108831Soundscape classification with convolutional neural networks reveals temporal and geographic patterns in ecoacoustic dataColin A. Quinn0Patrick Burns1Gurman Gill2Shrishail Baligar3Rose L. Snyder4Leonardo Salas5Scott J. Goetz6Matthew L. Clark7School of Informatics, Computing, and Cyber Systems, Northern Arizona University, Flagstaff, AZ, USA; Corresponding author at: School of Informatics, Computing, and Cyber Systems, Northern Arizona University, 1295 Knoles Dr, Flagstaff, AZ 86011, USA.School of Informatics, Computing, and Cyber Systems, Northern Arizona University, Flagstaff, AZ, USADepartment of Computer Science, Sonoma State University, Rohnert Park, CA, USAElectrical Engineering and Computer Science, University of California, Merced, CA, USAPoint Blue Conservation Science, Petaluma, CA, USAPoint Blue Conservation Science, Petaluma, CA, USASchool of Informatics, Computing, and Cyber Systems, Northern Arizona University, Flagstaff, AZ, USACenter for Interdisciplinary Geospatial Analysis, Geography, Environment and Planning, Sonoma State University, Rohnert Park, CA, USAInterest in ecoacoustics has resulted in an influx of acoustic data and novel methodologies to classify and relate landscape sound activity to biodiversity and ecosystem health. However, indicators used to summarize sound and quantify the effects of disturbances on biodiversity can be inconsistent when applied across ecological gradients. This study used an acoustic dataset of 487,148 min from 746 sites collected over 4 years across Sonoma County, California, USA, by citizen scientists. We built a custom labeled dataset of soundscape components and applied a deep learning framework to test our ability to predict these soundscape components: human noise (Anthropophony), wildlife vocalizations (Biophony), weather phenomena (Geophony), Quiet periods, and microphone Interference. These soundscape components allowed us to balance predicting variation in environmental recordings and relative time to build a custom labeled dataset. We used these data to quantify soundscape patterns across space and time that could be useful for environmental planning, ecosystem conservation and restoration, and biodiversity monitoring. We describe a pre-trained convolutional neural network, fine-tuned with our sound reference data, with classification achieving an overall F0.75-score of 0.88, precision of 0.94, and recall of 0.80 across the five target soundscape components. We deployed the model to predict soundscape components for all acoustic data and assess their hourly patterns. We noted an increase in Biophony in the early morning and evening, coinciding with peak animal community vocalization (e.g., dawn chorus). Anthropophony increased during morning/daylight hours and was lowest in the evenings, coinciding with diurnal patterns in human activity. Further, we examined soundscape patterns related to geographic properties at recording sites. Anthropophony decreased with increasing distance to major roads, while Quiet increased. Biophony and Quiet were comparable to Anthropophony at more urban/developed and agriculture/barren sites, while Biophony and Quiet were significantly higher than Anthropophony at less-developed shrubland, oak woodland, and conifer forest sites. These results demonstrate that acoustic classification of broad soundscape components is possible with small datasets, and classifications can be applied to a large acoustic dataset to gain ecological knowledge.http://www.sciencedirect.com/science/article/pii/S1470160X22003028Machine learningConvolutional neural network (CNN)EcoacousticsAnthropophonyBiophonyNaturally quiet landscapes
spellingShingle Colin A. Quinn
Patrick Burns
Gurman Gill
Shrishail Baligar
Rose L. Snyder
Leonardo Salas
Scott J. Goetz
Matthew L. Clark
Soundscape classification with convolutional neural networks reveals temporal and geographic patterns in ecoacoustic data
Ecological Indicators
Machine learning
Convolutional neural network (CNN)
Ecoacoustics
Anthropophony
Biophony
Naturally quiet landscapes
title Soundscape classification with convolutional neural networks reveals temporal and geographic patterns in ecoacoustic data
title_full Soundscape classification with convolutional neural networks reveals temporal and geographic patterns in ecoacoustic data
title_fullStr Soundscape classification with convolutional neural networks reveals temporal and geographic patterns in ecoacoustic data
title_full_unstemmed Soundscape classification with convolutional neural networks reveals temporal and geographic patterns in ecoacoustic data
title_short Soundscape classification with convolutional neural networks reveals temporal and geographic patterns in ecoacoustic data
title_sort soundscape classification with convolutional neural networks reveals temporal and geographic patterns in ecoacoustic data
topic Machine learning
Convolutional neural network (CNN)
Ecoacoustics
Anthropophony
Biophony
Naturally quiet landscapes
url http://www.sciencedirect.com/science/article/pii/S1470160X22003028
work_keys_str_mv AT colinaquinn soundscapeclassificationwithconvolutionalneuralnetworksrevealstemporalandgeographicpatternsinecoacousticdata
AT patrickburns soundscapeclassificationwithconvolutionalneuralnetworksrevealstemporalandgeographicpatternsinecoacousticdata
AT gurmangill soundscapeclassificationwithconvolutionalneuralnetworksrevealstemporalandgeographicpatternsinecoacousticdata
AT shrishailbaligar soundscapeclassificationwithconvolutionalneuralnetworksrevealstemporalandgeographicpatternsinecoacousticdata
AT roselsnyder soundscapeclassificationwithconvolutionalneuralnetworksrevealstemporalandgeographicpatternsinecoacousticdata
AT leonardosalas soundscapeclassificationwithconvolutionalneuralnetworksrevealstemporalandgeographicpatternsinecoacousticdata
AT scottjgoetz soundscapeclassificationwithconvolutionalneuralnetworksrevealstemporalandgeographicpatternsinecoacousticdata
AT matthewlclark soundscapeclassificationwithconvolutionalneuralnetworksrevealstemporalandgeographicpatternsinecoacousticdata