A wearable obstacle avoidance device for visually impaired individuals with cross-modal learning

31Citations
Citations of this article
41Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

It is challenging for wearable obstacle avoidance devices to simultaneously meet practical demands of high reliability, rapid response, long-lasting duration, and usable design. Here we report a wearable obstacle avoidance device, comprising a set of self-developed glasses (weighing ~400 grams, including an ~80 grams battery) and a common smartphone. Specifically, the glasses collect the multi-modal data for comprehensive environmental perception, including video and depth modalities, and implement a depth-aided video compression module. This module not only adaptively compresses video data to reduce transmission delay to the smartphone, but also operates on a customized FPGA board featuring a multi float-point vector unit streaming processing architecture, thereby facilitating responsive and energy-efficient obstacle detection. Additionally, we design a cross-modal obstacle detection module on the smartphone, which ensures reliable detection and provides user-friendly auditory and tactile alerts by utilizing cross-modal learning based on modal correlations. Multiple indoor and outdoor experimental results demonstrate 100% collision avoidance rates, delay of less than 320 ms, and duration of approximately 11 hours.

Cite

CITATION STYLE

APA

Gao, Y., Wu, D., Song, J., Zhang, X., Hou, B., Liu, H., … Zhou, L. (2025). A wearable obstacle avoidance device for visually impaired individuals with cross-modal learning. Nature Communications , 16(1). https://doi.org/10.1038/s41467-025-58085-x

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free