Mobility and Orientation Guidance for Individuals with Visual Impairments using AI
DOI:
https://doi.org/10.61453/INTIj.202571Keywords:
Internet of Things (IoT), Artificial Intelligence (AI), Bone Conducting, Haptic, Optical Character Recognition (OCR)Abstract
In this study, we developed an intelligent device and a smart application to improve the daily activities of the visually impaired individuals. Low-vision or blind people often face a number of barriers in the course of completing everyday tasks. Learning about roadways, purchasing commodities, reading written books, and digesting new information is significantly harder. To this end, a gadget was created to counter these obstacles. People with deficient eyesight or complete blindness can now enjoy the effect of reading books and articles in real-time using OCR and AI-powered technology. They can also recognize things, goods, and people, including visual information like facial expressions. In addition, haptic feedback through bone-conducting headphones gives multilingual notifications of either vehicle movement or road condition.
References
Al-Shamma, O., Fathi, S., & Al-Samarraie, M. Q. (2023). A smart glasses-based navigation system for the visually impaired using deep learning and IoT. IEEE Sensors Journal, *23*(5), 5123-5132.
Bai, J., Liu, Z., & Li, Y. (2021). An integrated wearable system for outdoor navigation and obstacle recognition for the visually impaired. ACM Transactions on Accessible Computing, *14*(2), 1-25. https://doi.org/10.3390/electronics8060697
Gokul, K., Hariharan, S., & S., M. (2024, December 12). Smart indoor navigation device for blind people with obstacle and staircase detection. 2024 4th International Conference on Ubiquitous Computing and Intelligent Information Systems (ICUIS), 970–978. https://doi.org/10.1109/ICUIS64676.2024.10867178
Huu-Huy, N., Hung Linh, L., & Feng-Cheng, L. (2025, May 23). Deep-learning-based cognitive assistance embedded systems for people with visual impairment. Applied Sciences, 15(11), 5887. https://doi.org/10.3390/app15115887
Isma, H., Sara, M., & Qamar, N. (2023, October 29). Realtime object detection and recognition device for blind and partially blind people (NANOEYE). 2023 Computer Applications & Technological Solutions (CATS), 1–7. https://doi.org/10.1109/CATS58689.2023.00010
Jong Hyeok, H., Inkwon, Y., Hyun Soo, K. B. J., Ji Hwan, M., Jinseok, P., & Hee-Jae, J. (2024, October 24). Mobility support with intelligent obstacle detection for enhanced safety. Optics, 5(4), 434–444. https://doi.org/10.3390/opt5040025
Kim, T., Park, S., & Lee, J. (2022). A haptic-feedback smart cane system integrating computer vision and sensor fusion. Sensors and Actuators A: Physical, *344*, 113733.
Lee, A., & Wang, C. (2024). Enhancing public transit accessibility through real-time auditory guidance systems. International Journal of Human–Computer Interaction, *40*(1), 45-59.
Mozibul, H., Monir, H., Md. Yeasin, A., Nazmun Nessa, M., Asma, M., & Mohammad Raiful, I. (2021, June 29). An AI-based navigation aid for the visually impaired. 2021 International Conference on Artificial Intelligence and Computer Science Technology (ICAICST), 217–222. https://doi.org/10.1109/ICAICST53116.2021.9497799
Patel, R., & Kumar, S. (2023). Edge AI for low-latency obstacle avoidance in wearable assistive devices. IEEE Transactions on Medical Robotics and Bionics, *5*(3), 678-689
RICHARD, K., PATRIK, J., & TIBOR, V. (2025, January). Low-cost multifunctional assistive device for visually impaired individuals. Acta Technologica Dubnicae, 15(1), 1–15. https://doi.org/10.1109/ACCESS.2025.3554366
Smith, J. A., & Chen, L. (2022). Auditory interfaces for wearable navigation aids: A user-centered design review. ACM Computing Surveys, *55*(8), 1-38.
Stephanie, M., Valentina, K., Victoria, T., & Cristina, M. (2022). Assisted mobility: An application to help the blind to use public transport. 2022 IEEE Global Humanitarian Technology Conference (GHTC), 445–448. https://doi.org/10.1109/GHTC55712.2022.9911013
Vinish, P., Karthik, M., & Divya, P. (2024, June 1). Assistive systems for visually impaired persons: Challenges and opportunities for navigation assistance. *2024 2nd DMIHER International Conference on Artificial Intelligence in Healthcare, Education and Industry (IDICAIEI), 24*(11), 3572. https://doi.org/10.3390/s24113572
Xiaochen, Z., Xiaoyu, H., Yiran, D., Liumei, L., Wujing, L., & Xing, X. (2024). Advancements in smart wearable mobility aids for visual impairments: A bibliometric narrative review. Sensors, 24(23), 7986. https://doi.org/10.3390/s24247986
Zhang, Y., & Rodriguez, A. (2023). Cross-modal sensory integration in electronic travel aids: From theory to practice. IEEE Reviews in Biomedical Engineering, *16*, 432-447.
Zhao, Q., & Ito, M. (2024). Evaluation of user experience and trust in autonomous navigation assistants for the blind. Universal Access in the Information Society, *23*, 1257-1273.
Zou, B., Mendez, O., & Gupta, S. (2025). Federated learning for privacy-preserving navigation data in assistive technology ecosystems. Pervasive and Mobile Computing, 98, 101872.
Downloads
Published
How to Cite
Issue
Section
License
Copyright (c) 2025 INTI Journal

This work is licensed under a Creative Commons Attribution 4.0 International License.