A texture rendering system relying on pseudo-haptic and audio feedback is presented in this paper. While the user touches the texture displayed on a tactile screen, the associated image is deformed according to the contact area and the rubbing motion to simulate pressure. Additionally audio feedback is synthesized in real-time to simulate friction. A novel example-based scheme takes advantage of recorded audio samples of friction between actual textures and a finger at several speeds to synthesize the final output sound. This system can be implemented on any existing tactile screen without any extra mechanical device.
CITATION STYLE
Fleureau, J., Lefevre, Y., Danieau, F., Guillotel, P., & Costes, A. (2016). Texture rendering on a tactile surface using extended elastic images and example-based audio cues. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 9774, pp. 350–359). Springer Verlag. https://doi.org/10.1007/978-3-319-42321-0_32
Mendeley helps you to discover research relevant for your work.