The HandSight project investigates how wearable micro-cameras can be used to augment a blind or visually impaired user's sense of touch with computer vision. Our goal is to support an array of activities of daily living by sensing and feeding back non-tactile information (e.g., color, printed text, patterns) about an object as it is touched. In this poster paper, we provide an overview of the project, our current proof-of-concept prototype, and a summary of findings from finger-based text reading studies. As this is an early-stage project, we also enumerate current open questions.
CITATION STYLE
Findlater, L., Stearns, L., Du, R., Oh, U., Ross, D., Chellappa, R., & Froehlich, J. E. (2015). Supporting everyday activities for persons with visual impairments through computer vision-augmented touch. In ASSETS 2015 - Proceedings of the 17th International ACM SIGACCESS Conference on Computers and Accessibility (pp. 383–384). Association for Computing Machinery, Inc. https://doi.org/10.1145/2700648.2811381
Mendeley helps you to discover research relevant for your work.