Environmental acoustic transformation and feature extraction for machine hearing

3Citations
Citations of this article
6Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

This paper explores the transformation of environmental sound waveform and feature set into a parametric type representation to be used in analysis, recognition, and identification for auditory analysis of machine hearing systems. Generally, the focus of the research and study in sound recognition is concentrated on the music and speech domains, on the other hand, there are limited in non-speech environmental recognition. We analyzed and evaluated the different current feature algorithms and methods explored for the acoustic recognition of environmental sounds, the Mel Filterbank Energies (FBEs) and Gammatone spectral coefficients (GSTC) and for classifying the sound signal the Convolutional Neural Network (CNN) was used. The result shows that GSTC performs well as a feature compared to FBEs, but FBEs tend to perform better when combined with other feature. This shows that a combination of features set is promising in obtaining a higher accuracy compared to a single feature in environmental sound classification, that is helpful in the development of the machine hearing systems.

Cite

CITATION STYLE

APA

Catanghal, R., Palaoag, T., & Dayagdag, C. (2019). Environmental acoustic transformation and feature extraction for machine hearing. In IOP Conference Series: Materials Science and Engineering (Vol. 482). Institute of Physics Publishing. https://doi.org/10.1088/1757-899X/482/1/012007

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free