Many perceptual models for audio reconstruction have been proposed to create the virtual sound, but the direction of the virtual sound maybe deviate from the desired direction due to the distortion of binaural cues. In this paper, a binaural cues’ equation for real sound and virtual one reproduced by dual loudspeakers is established to derive weight vectors based on the head-related transfer function (HRTF). After being filtered by the weight vectors, sound signals emitted from the loudspeakers can deliver an accurate spatial impression to the listener. However, the HRTFs change with listeners, by which the weight vectors calculated also vary from person to person. Therefore, a radial basis function neural network (RBFNN) is designed to personalize weight vectors for each specific listener. Compared with the three methods including vector base amplitude panning (VBAP), the HRTF-based panning (HP) and the band-based panning (BP), the method in this paper can reproduce binaural cues more accurately, and subjective test also indicates that there is no significant difference in perception between real sound and virtual sound based on the proposed methods.
CITATION STYLE
Zheng, J., Tu, W., Zhang, X., Yang, W., Zhai, S., & Shen, C. (2018). A sound image reproduction model based on personalized weight vectors. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 11165 LNCS, pp. 607–617). Springer Verlag. https://doi.org/10.1007/978-3-030-00767-6_56
Mendeley helps you to discover research relevant for your work.