VGGish-based detection of biological sound components and their spatio-temporal variations in a subtropical forest in eastern China

1Citations
Citations of this article
10Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Passive acoustic monitoring technology is widely used to monitor the diversity of vocal animals, but the question of how to quickly extract effective sound patterns remains a challenge due to the difficulty of distinguishing biological sounds within multiple sound sources in a soundscape. In this study, we address the potential application of the VGGish model, pre-trained on Google's AudioSet dataset, for the extraction of acoustic features, together with an unsupervised clustering method based on the Gaussian mixture model, to identify various sound sources from a soundscape of a subtropical forest in China. The results show that different biotic and abiotic components can be distinguished from various confounding sound sources. Birds and insects were the two primary biophony sound sources, and their sounds displayed distinct temporal patterns across both diurnal and monthly time frames and distinct spatial patterns in the landscape. Using the clustering and modeling method of the general sound feature set, we quickly depicted the soundscape in a subtropical forest ecosystem, which could be used to track dynamic changes in the acoustic environment and provide help for biodiversity and ecological environment monitoring.

Cite

CITATION STYLE

APA

Wang, M., Mei, J., Darras, K. F., & Liu, F. (2023). VGGish-based detection of biological sound components and their spatio-temporal variations in a subtropical forest in eastern China. PeerJ, 11. https://doi.org/10.7717/peerj.16462

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free