Sound signal invariant DAE neural network-based quantizer architecture of audio/speech coder using the matching pursuit algorithm

1Citations
Citations of this article
3Readers
Mendeley users who have this article in their library.
Get full text

Abstract

The paper is devoted to the quantization algorithm development based on the neural networks framework. This research is considered in the context of the scalable real-time audio/speech coder based on the perceptually adaptive matching pursuit algorithm. The encoder parameterizes the input sound signal frame with some amount of real numbers that are need to be compactly represented in binary form, i.e. quantized. The neural network quantization approach gives great opportunity for such a goal because the data quantized in whole vector but not in separate form and it can effectively use correlations between each element of the input coded vector. Deep autoencoder (DAE) neural network-based architecture for the quantization part of the encoding algorithm is shown. Its structure and learning features are described. Conducted experiments points out the big compression ratio with high reconstructed signal quality of the developed audio/speech coder quantization scheme.

Cite

CITATION STYLE

APA

Avramov, V., Herasimovich, V., & Petrovsky, A. (2018). Sound signal invariant DAE neural network-based quantizer architecture of audio/speech coder using the matching pursuit algorithm. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 10878 LNCS, pp. 511–520). Springer Verlag. https://doi.org/10.1007/978-3-319-92537-0_59

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free