Soundscape classification with convolutional neural networks reveals temporal and geographic patterns in ecoacoustic data
Interest in ecoacoustics has resulted in an influx of acoustic data and novel methodologies to classify and relate landscape sound activity to biodiversity and ecosystem health. However, indicators used to summarize sound and quantify the effects of disturbances on biodiversity can be inconsistent w...
Main Authors: | , , , , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Elsevier
2022-05-01
|
Series: | Ecological Indicators |
Subjects: | |
Online Access: | http://www.sciencedirect.com/science/article/pii/S1470160X22003028 |
_version_ | 1818010008569249792 |
---|---|
author | Colin A. Quinn Patrick Burns Gurman Gill Shrishail Baligar Rose L. Snyder Leonardo Salas Scott J. Goetz Matthew L. Clark |
author_facet | Colin A. Quinn Patrick Burns Gurman Gill Shrishail Baligar Rose L. Snyder Leonardo Salas Scott J. Goetz Matthew L. Clark |
author_sort | Colin A. Quinn |
collection | DOAJ |
description | Interest in ecoacoustics has resulted in an influx of acoustic data and novel methodologies to classify and relate landscape sound activity to biodiversity and ecosystem health. However, indicators used to summarize sound and quantify the effects of disturbances on biodiversity can be inconsistent when applied across ecological gradients. This study used an acoustic dataset of 487,148 min from 746 sites collected over 4 years across Sonoma County, California, USA, by citizen scientists. We built a custom labeled dataset of soundscape components and applied a deep learning framework to test our ability to predict these soundscape components: human noise (Anthropophony), wildlife vocalizations (Biophony), weather phenomena (Geophony), Quiet periods, and microphone Interference. These soundscape components allowed us to balance predicting variation in environmental recordings and relative time to build a custom labeled dataset. We used these data to quantify soundscape patterns across space and time that could be useful for environmental planning, ecosystem conservation and restoration, and biodiversity monitoring. We describe a pre-trained convolutional neural network, fine-tuned with our sound reference data, with classification achieving an overall F0.75-score of 0.88, precision of 0.94, and recall of 0.80 across the five target soundscape components. We deployed the model to predict soundscape components for all acoustic data and assess their hourly patterns. We noted an increase in Biophony in the early morning and evening, coinciding with peak animal community vocalization (e.g., dawn chorus). Anthropophony increased during morning/daylight hours and was lowest in the evenings, coinciding with diurnal patterns in human activity. Further, we examined soundscape patterns related to geographic properties at recording sites. Anthropophony decreased with increasing distance to major roads, while Quiet increased. Biophony and Quiet were comparable to Anthropophony at more urban/developed and agriculture/barren sites, while Biophony and Quiet were significantly higher than Anthropophony at less-developed shrubland, oak woodland, and conifer forest sites. These results demonstrate that acoustic classification of broad soundscape components is possible with small datasets, and classifications can be applied to a large acoustic dataset to gain ecological knowledge. |
first_indexed | 2024-04-14T05:49:58Z |
format | Article |
id | doaj.art-eaeda47cfe52422385952dbf25142ecb |
institution | Directory Open Access Journal |
issn | 1470-160X |
language | English |
last_indexed | 2024-04-14T05:49:58Z |
publishDate | 2022-05-01 |
publisher | Elsevier |
record_format | Article |
series | Ecological Indicators |
spelling | doaj.art-eaeda47cfe52422385952dbf25142ecb2022-12-22T02:09:09ZengElsevierEcological Indicators1470-160X2022-05-01138108831Soundscape classification with convolutional neural networks reveals temporal and geographic patterns in ecoacoustic dataColin A. Quinn0Patrick Burns1Gurman Gill2Shrishail Baligar3Rose L. Snyder4Leonardo Salas5Scott J. Goetz6Matthew L. Clark7School of Informatics, Computing, and Cyber Systems, Northern Arizona University, Flagstaff, AZ, USA; Corresponding author at: School of Informatics, Computing, and Cyber Systems, Northern Arizona University, 1295 Knoles Dr, Flagstaff, AZ 86011, USA.School of Informatics, Computing, and Cyber Systems, Northern Arizona University, Flagstaff, AZ, USADepartment of Computer Science, Sonoma State University, Rohnert Park, CA, USAElectrical Engineering and Computer Science, University of California, Merced, CA, USAPoint Blue Conservation Science, Petaluma, CA, USAPoint Blue Conservation Science, Petaluma, CA, USASchool of Informatics, Computing, and Cyber Systems, Northern Arizona University, Flagstaff, AZ, USACenter for Interdisciplinary Geospatial Analysis, Geography, Environment and Planning, Sonoma State University, Rohnert Park, CA, USAInterest in ecoacoustics has resulted in an influx of acoustic data and novel methodologies to classify and relate landscape sound activity to biodiversity and ecosystem health. However, indicators used to summarize sound and quantify the effects of disturbances on biodiversity can be inconsistent when applied across ecological gradients. This study used an acoustic dataset of 487,148 min from 746 sites collected over 4 years across Sonoma County, California, USA, by citizen scientists. We built a custom labeled dataset of soundscape components and applied a deep learning framework to test our ability to predict these soundscape components: human noise (Anthropophony), wildlife vocalizations (Biophony), weather phenomena (Geophony), Quiet periods, and microphone Interference. These soundscape components allowed us to balance predicting variation in environmental recordings and relative time to build a custom labeled dataset. We used these data to quantify soundscape patterns across space and time that could be useful for environmental planning, ecosystem conservation and restoration, and biodiversity monitoring. We describe a pre-trained convolutional neural network, fine-tuned with our sound reference data, with classification achieving an overall F0.75-score of 0.88, precision of 0.94, and recall of 0.80 across the five target soundscape components. We deployed the model to predict soundscape components for all acoustic data and assess their hourly patterns. We noted an increase in Biophony in the early morning and evening, coinciding with peak animal community vocalization (e.g., dawn chorus). Anthropophony increased during morning/daylight hours and was lowest in the evenings, coinciding with diurnal patterns in human activity. Further, we examined soundscape patterns related to geographic properties at recording sites. Anthropophony decreased with increasing distance to major roads, while Quiet increased. Biophony and Quiet were comparable to Anthropophony at more urban/developed and agriculture/barren sites, while Biophony and Quiet were significantly higher than Anthropophony at less-developed shrubland, oak woodland, and conifer forest sites. These results demonstrate that acoustic classification of broad soundscape components is possible with small datasets, and classifications can be applied to a large acoustic dataset to gain ecological knowledge.http://www.sciencedirect.com/science/article/pii/S1470160X22003028Machine learningConvolutional neural network (CNN)EcoacousticsAnthropophonyBiophonyNaturally quiet landscapes |
spellingShingle | Colin A. Quinn Patrick Burns Gurman Gill Shrishail Baligar Rose L. Snyder Leonardo Salas Scott J. Goetz Matthew L. Clark Soundscape classification with convolutional neural networks reveals temporal and geographic patterns in ecoacoustic data Ecological Indicators Machine learning Convolutional neural network (CNN) Ecoacoustics Anthropophony Biophony Naturally quiet landscapes |
title | Soundscape classification with convolutional neural networks reveals temporal and geographic patterns in ecoacoustic data |
title_full | Soundscape classification with convolutional neural networks reveals temporal and geographic patterns in ecoacoustic data |
title_fullStr | Soundscape classification with convolutional neural networks reveals temporal and geographic patterns in ecoacoustic data |
title_full_unstemmed | Soundscape classification with convolutional neural networks reveals temporal and geographic patterns in ecoacoustic data |
title_short | Soundscape classification with convolutional neural networks reveals temporal and geographic patterns in ecoacoustic data |
title_sort | soundscape classification with convolutional neural networks reveals temporal and geographic patterns in ecoacoustic data |
topic | Machine learning Convolutional neural network (CNN) Ecoacoustics Anthropophony Biophony Naturally quiet landscapes |
url | http://www.sciencedirect.com/science/article/pii/S1470160X22003028 |
work_keys_str_mv | AT colinaquinn soundscapeclassificationwithconvolutionalneuralnetworksrevealstemporalandgeographicpatternsinecoacousticdata AT patrickburns soundscapeclassificationwithconvolutionalneuralnetworksrevealstemporalandgeographicpatternsinecoacousticdata AT gurmangill soundscapeclassificationwithconvolutionalneuralnetworksrevealstemporalandgeographicpatternsinecoacousticdata AT shrishailbaligar soundscapeclassificationwithconvolutionalneuralnetworksrevealstemporalandgeographicpatternsinecoacousticdata AT roselsnyder soundscapeclassificationwithconvolutionalneuralnetworksrevealstemporalandgeographicpatternsinecoacousticdata AT leonardosalas soundscapeclassificationwithconvolutionalneuralnetworksrevealstemporalandgeographicpatternsinecoacousticdata AT scottjgoetz soundscapeclassificationwithconvolutionalneuralnetworksrevealstemporalandgeographicpatternsinecoacousticdata AT matthewlclark soundscapeclassificationwithconvolutionalneuralnetworksrevealstemporalandgeographicpatternsinecoacousticdata |