Mapping phonetic features for voice-driven sound synthesis

3Citations
Citations of this article
6Readers
Mendeley users who have this article in their library.
Get full text

Abstract

In applications where the human voice controls the synthesis of musical instruments sounds, phonetics convey musical information that might be related to the sound of the imitated musical instrument. Our initial hypothesis is that phonetics are user- and instrument-dependent, but they remain constant for a single subject and instrument. We propose a user-adapted system, where mappings from voice features to synthesis parameters depend on how subjects sing musical articulations, i.e. note to note transitions. The system consists of two components. First, a voice signal segmentation module that automatically determines note-to-note transitions. Second, a classifier that determines the type of musical articulation for each transition based on a set of phonetic features. For validating our hypothesis, we run an experiment where subjects imitated real instrument recordings with their voice. Performance recordings consisted of short phrases of saxophone and violin performed in three grades of musical articulation labeled as: staccato, normal, legato. The results of a supervised training classifier (user-dependent) are compared to a classifier based on heuristic rules (user-independent). Finally, from the previous results we show how to control the articulation in a sample-concatenation synthesizer by selecting the most appropriate samples. © 2008 Springer-Verlag.

Cite

CITATION STYLE

APA

Janer, J., & Maestre, E. (2008). Mapping phonetic features for voice-driven sound synthesis. In Communications in Computer and Information Science (Vol. 23 CCIS, pp. 304–314). Springer Verlag. https://doi.org/10.1007/978-3-540-88653-2_23

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free