A sound image reproduction model based on personalized weight vectors

1Citations
Citations of this article
1Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Many perceptual models for audio reconstruction have been proposed to create the virtual sound, but the direction of the virtual sound maybe deviate from the desired direction due to the distortion of binaural cues. In this paper, a binaural cues’ equation for real sound and virtual one reproduced by dual loudspeakers is established to derive weight vectors based on the head-related transfer function (HRTF). After being filtered by the weight vectors, sound signals emitted from the loudspeakers can deliver an accurate spatial impression to the listener. However, the HRTFs change with listeners, by which the weight vectors calculated also vary from person to person. Therefore, a radial basis function neural network (RBFNN) is designed to personalize weight vectors for each specific listener. Compared with the three methods including vector base amplitude panning (VBAP), the HRTF-based panning (HP) and the band-based panning (BP), the method in this paper can reproduce binaural cues more accurately, and subjective test also indicates that there is no significant difference in perception between real sound and virtual sound based on the proposed methods.

Cite

CITATION STYLE

APA

Zheng, J., Tu, W., Zhang, X., Yang, W., Zhai, S., & Shen, C. (2018). A sound image reproduction model based on personalized weight vectors. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 11165 LNCS, pp. 607–617). Springer Verlag. https://doi.org/10.1007/978-3-030-00767-6_56

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free