Authors :
Thanh Tuan Ton That; Khang Huy Ngo; Nam Hien Cao Nguyen
Volume/Issue :
Volume 11 - 2026, Issue 2 - February
Google Scholar :
https://tinyurl.com/yauw3yy8
Scribd :
https://tinyurl.com/36kf42kf
DOI :
https://doi.org/10.38124/ijisrt/26feb637
Note : A published paper may take 4-5 working days from the publication date to appear in PlumX Metrics, Semantic Scholar, and ResearchGate.
Abstract :
Visual impairment and age-related decline significantly affect independent mobility and safety in daily life. This
paper presents a smart monitoring and assistance system aimed at supporting visually impaired individuals and elderly people
through the integrating of artificial intelligence, computer vision, and sensor-based technologies. Getting inspired by the
working mechanism of the human eye, the proposed system employs the YOLO11 deep learning model for real-time object
detection and classification, combined with the Depth Anything v2 model for monocular depth estimation to calculate the
distance between users and surrounding objects.
The system is implemented using an embedded camera and IoT-based sensors, including ultrasonic distance sensors, GNSS
positioning, heart rate monitoring (MAX30102), and fall detection modules, enabling comprehensive environmental perception
and user health monitoring. Experimental evaluations were conducted in both bright and low-light environments using a self-
collected dataset. The results depict that the proposed system achieves an overall the accurate of detection – approximately 95
percent, with stable performance across varying lighting conditions.
The findings confirm the feasibility and effectiveness of integrating deep learning models with embedded hardware to
provide real-time assistance. This system has strong potential for development into a wearable smart device capable of enhancing
mobility, reducing collision risks, and improving the independence and quality of life for visually impaired and elderly users.
Moreover, the proposed approach contributes to the advancement of human-centered intelligent assistive technologies with
meaningful social impact.
Keywords :
Assistive Technology, Visual Impairment, Elderly Care, Object Detection, Depth Estimation, YOLO, Computer Vision, Internet of Things (IoT).
References :
- Jan Egger, Christina Gsaxner, Xiaojun Chen, Jiang Bian, Jens Kleesiek, and Behrus Puladi. Apple vision pro for healthcare:” the ultimate display. arXiv preprint arXiv: 2308.04313, (2), 2023.
- Seth R Flaxman, Rupert RA Bourne, Serge Resnikoff, Peter Ackland, Tasanee Braithwaite, Maria V Cicinelli, Aditi Das, Jost B Jonas, Jill Keeffe, John H Kempen, et al. Global causes of blindness and distance vision impairment 1990–2020: a systematic review and meta-analysis. The Lancet Global Health, 5(12):e1221–e1234, 2017.
- JASD Fonseca, Antonio Baptista, Ma Joao Martins, and Joao Paulo N Torres. Distance measurement systems using lasers and their applications. Applied Physics Research, 9(4):33–43, 2017.
- Tung Sum Thomas Kwok, Zeyong Zhang, Chi-Hua Wang, and Guang Cheng. Towards high supervised learning utility training data generation: Data pruning and column reordering. arXiv preprint arXiv:2507.10088, 2025.
- Dengsheng Lu and Qihao Weng. A survey of image classification methods and techniques for improving classification performance. International journal of Remote sensing, 28(5):823–870, 2007.
- Michael Moll. Displacement damage in silicon detectors for high energy physics. IEEE Transactions on Nuclear Science, 65(8):1561–1582, 2018.
- Nitin Rane. Yolo and faster r-cnn object detection for smart industry 4.0 and industry 5.0: applications, challenges, and opportunities. Available at SSRN 4624206, 2023.
- Joseph Redmon, Santosh Divvala, Ross Girshick, and Ali Farhadi. You only look once: Unified, real-time object detection. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 779–788, 2016.
- Hector Rodr´ ´ıguez-Rangel, Luis Alberto Morales-Rosales, Rafael Imperial-Rojo, Mario Alberto Roman-Garay, Gloria Ekaterine Peralta-Penu˜ nuri, and Mariana Lobato-B˜ aez. Analysis´ of statistical and artificial intelligence algorithms for real-time speed estimation based on vehicle detection with yolo. Applied Sciences, 12(6):2907, 2022.
- Manoj Sahni, Ritu Sahni, and Jose M Merig´ o.´ Neural Networks, Machine Learning, and Image Processing: Mathematical Modeling and Applications. CRC Press, 2022.
- Mohammad Javad Shafiee, Brendan Chywl, Francis Li, and Alexander Wong. Fast yolo: A fast you only look once system for real-time embedded object detection in video. arXiv preprint arXiv:1709.05943, 2017.
- Akhilesh Sharma, Vipan Kumar, and Louis Longchamps. Comparative performance of yolov8, yolov9, yolov10, yolov11 and faster r-cnn models for detection of multiple weed species. Smart Agricultural Technology, 9:100648, 2024.
- Do Thuan. Evolution of yolo algorithm and yolov5: The state-of-the-art object detention algorithm. 2021.
- Tao Xu, Wei Shen, Xiaoshan Lin, and Yi Min Xie. Mechanical properties of additively manufactured thermoplastic polyurethane (tpu) material affected by various processing parameters. Polymers, 12(12):3010, 2020.
- Wei You, Changqing Shen, Dong Wang, Liang Chen, Xingxing Jiang, and Zhongkui Zhu. An intelligent deep feature learning method with improved activation functions for machine fault diagnosis. IEEE access, 8:1975–1985, 2019.
- Xing Zhang, Gongjian Wen, and Wei Dai. A tensor decomposition-based anomaly detection algorithm for hyperspectral image. IEEE Transactions on Geoscience and Remote Sensing, 54(10):5801–5820, 2016.
Visual impairment and age-related decline significantly affect independent mobility and safety in daily life. This
paper presents a smart monitoring and assistance system aimed at supporting visually impaired individuals and elderly people
through the integrating of artificial intelligence, computer vision, and sensor-based technologies. Getting inspired by the
working mechanism of the human eye, the proposed system employs the YOLO11 deep learning model for real-time object
detection and classification, combined with the Depth Anything v2 model for monocular depth estimation to calculate the
distance between users and surrounding objects.
The system is implemented using an embedded camera and IoT-based sensors, including ultrasonic distance sensors, GNSS
positioning, heart rate monitoring (MAX30102), and fall detection modules, enabling comprehensive environmental perception
and user health monitoring. Experimental evaluations were conducted in both bright and low-light environments using a self-
collected dataset. The results depict that the proposed system achieves an overall the accurate of detection – approximately 95
percent, with stable performance across varying lighting conditions.
The findings confirm the feasibility and effectiveness of integrating deep learning models with embedded hardware to
provide real-time assistance. This system has strong potential for development into a wearable smart device capable of enhancing
mobility, reducing collision risks, and improving the independence and quality of life for visually impaired and elderly users.
Moreover, the proposed approach contributes to the advancement of human-centered intelligent assistive technologies with
meaningful social impact.
Keywords :
Assistive Technology, Visual Impairment, Elderly Care, Object Detection, Depth Estimation, YOLO, Computer Vision, Internet of Things (IoT).