An Intelligent Assistive and Monitoring System for Elderly and Visually Impaired Users


Authors : Thanh Tuan Ton That; Khang Huy Ngo; Nam Hien Cao Nguyen

Volume/Issue : Volume 11 - 2026, Issue 2 - February


Google Scholar : https://tinyurl.com/yauw3yy8

Scribd : https://tinyurl.com/36kf42kf

DOI : https://doi.org/10.38124/ijisrt/26feb637

Note : A published paper may take 4-5 working days from the publication date to appear in PlumX Metrics, Semantic Scholar, and ResearchGate.


Abstract : Visual impairment and age-related decline significantly affect independent mobility and safety in daily life. This paper presents a smart monitoring and assistance system aimed at supporting visually impaired individuals and elderly people through the integrating of artificial intelligence, computer vision, and sensor-based technologies. Getting inspired by the working mechanism of the human eye, the proposed system employs the YOLO11 deep learning model for real-time object detection and classification, combined with the Depth Anything v2 model for monocular depth estimation to calculate the distance between users and surrounding objects. The system is implemented using an embedded camera and IoT-based sensors, including ultrasonic distance sensors, GNSS positioning, heart rate monitoring (MAX30102), and fall detection modules, enabling comprehensive environmental perception and user health monitoring. Experimental evaluations were conducted in both bright and low-light environments using a self- collected dataset. The results depict that the proposed system achieves an overall the accurate of detection – approximately 95 percent, with stable performance across varying lighting conditions. The findings confirm the feasibility and effectiveness of integrating deep learning models with embedded hardware to provide real-time assistance. This system has strong potential for development into a wearable smart device capable of enhancing mobility, reducing collision risks, and improving the independence and quality of life for visually impaired and elderly users. Moreover, the proposed approach contributes to the advancement of human-centered intelligent assistive technologies with meaningful social impact.

Keywords : Assistive Technology, Visual Impairment, Elderly Care, Object Detection, Depth Estimation, YOLO, Computer Vision, Internet of Things (IoT).

References :

  1. Jan Egger, Christina Gsaxner, Xiaojun Chen, Jiang Bian, Jens Kleesiek, and Behrus Puladi. Apple vision pro for healthcare:” the ultimate display. arXiv preprint arXiv: 2308.04313, (2), 2023.
  2. Seth R Flaxman, Rupert RA Bourne, Serge Resnikoff, Peter Ackland, Tasanee Braithwaite, Maria V Cicinelli, Aditi Das, Jost B Jonas, Jill Keeffe, John H Kempen, et al. Global causes of blindness and distance vision impairment 1990–2020: a systematic review and meta-analysis. The Lancet Global Health, 5(12):e1221–e1234, 2017.
  3. JASD Fonseca, Antonio Baptista, Ma Joao Martins, and Joao Paulo N Torres. Distance measurement systems using lasers and their applications. Applied Physics Research, 9(4):33–43, 2017.
  4. Tung Sum Thomas Kwok, Zeyong Zhang, Chi-Hua Wang, and Guang Cheng. Towards high supervised learning utility training data generation: Data pruning and column reordering. arXiv preprint arXiv:2507.10088, 2025.
  5. Dengsheng Lu and Qihao Weng. A survey of image classification methods and techniques for improving classification performance. International journal of Remote sensing, 28(5):823–870, 2007.
  6. Michael Moll. Displacement damage in silicon detectors for high energy physics. IEEE Transactions on Nuclear Science, 65(8):1561–1582, 2018.
  7. Nitin Rane. Yolo and faster r-cnn object detection for smart industry 4.0 and industry 5.0: applications, challenges, and opportunities. Available at SSRN 4624206, 2023.
  8. Joseph Redmon, Santosh Divvala, Ross Girshick, and Ali Farhadi. You only look once: Unified, real-time object detection. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 779–788, 2016.
  9. Hector Rodr´ ´ıguez-Rangel, Luis Alberto Morales-Rosales, Rafael Imperial-Rojo, Mario Alberto Roman-Garay, Gloria Ekaterine Peralta-Penu˜ nuri, and Mariana Lobato-B˜ aez. Analysis´ of statistical and artificial intelligence algorithms for real-time speed estimation based on vehicle detection with yolo. Applied Sciences, 12(6):2907, 2022.
  10. Manoj Sahni, Ritu Sahni, and Jose M Merig´ o.´ Neural Networks, Machine Learning, and Image Processing: Mathematical Modeling and Applications. CRC Press, 2022.
  11. Mohammad Javad Shafiee, Brendan Chywl, Francis Li, and Alexander Wong. Fast yolo: A fast you only look once system for real-time embedded object detection in video. arXiv preprint arXiv:1709.05943, 2017.
  12. Akhilesh Sharma, Vipan Kumar, and Louis Longchamps. Comparative performance of yolov8, yolov9, yolov10, yolov11 and faster r-cnn models for detection of multiple weed species. Smart Agricultural Technology, 9:100648, 2024.
  13. Do Thuan. Evolution of yolo algorithm and yolov5: The state-of-the-art object detention algorithm. 2021.
  14. Tao Xu, Wei Shen, Xiaoshan Lin, and Yi Min Xie. Mechanical properties of additively manufactured thermoplastic polyurethane (tpu) material affected by various processing parameters. Polymers, 12(12):3010, 2020.
  15. Wei You, Changqing Shen, Dong Wang, Liang Chen, Xingxing Jiang, and Zhongkui Zhu. An intelligent deep feature learning method with improved activation functions for machine fault diagnosis. IEEE access, 8:1975–1985, 2019.
  16. Xing Zhang, Gongjian Wen, and Wei Dai. A tensor decomposition-based anomaly detection algorithm for hyperspectral image. IEEE Transactions on Geoscience and Remote Sensing, 54(10):5801–5820, 2016.

Visual impairment and age-related decline significantly affect independent mobility and safety in daily life. This paper presents a smart monitoring and assistance system aimed at supporting visually impaired individuals and elderly people through the integrating of artificial intelligence, computer vision, and sensor-based technologies. Getting inspired by the working mechanism of the human eye, the proposed system employs the YOLO11 deep learning model for real-time object detection and classification, combined with the Depth Anything v2 model for monocular depth estimation to calculate the distance between users and surrounding objects. The system is implemented using an embedded camera and IoT-based sensors, including ultrasonic distance sensors, GNSS positioning, heart rate monitoring (MAX30102), and fall detection modules, enabling comprehensive environmental perception and user health monitoring. Experimental evaluations were conducted in both bright and low-light environments using a self- collected dataset. The results depict that the proposed system achieves an overall the accurate of detection – approximately 95 percent, with stable performance across varying lighting conditions. The findings confirm the feasibility and effectiveness of integrating deep learning models with embedded hardware to provide real-time assistance. This system has strong potential for development into a wearable smart device capable of enhancing mobility, reducing collision risks, and improving the independence and quality of life for visually impaired and elderly users. Moreover, the proposed approach contributes to the advancement of human-centered intelligent assistive technologies with meaningful social impact.

Keywords : Assistive Technology, Visual Impairment, Elderly Care, Object Detection, Depth Estimation, YOLO, Computer Vision, Internet of Things (IoT).

Paper Submission Last Date
31 - March - 2026

SUBMIT YOUR PAPER CALL FOR PAPERS
Video Explanation for Published paper

Never miss an update from Papermashup

Get notified about the latest tutorials and downloads.

Subscribe by Email

Get alerts directly into your inbox after each post and stay updated.
Subscribe
OR

Subscribe by RSS

Add our RSS to your feedreader to get regular updates from us.
Subscribe