Authors :
Saniya Pathan; Swaraj Nangare; Riya Patil; Sakshi Renuse; Anagha Chaphadkar
Volume/Issue :
Volume 11 - 2026, Issue 4 - April
Google Scholar :
https://tinyurl.com/4s2bjua3
Scribd :
https://tinyurl.com/4cfn52au
DOI :
https://doi.org/10.38124/ijisrt/26apr1716
Note : A published paper may take 4-5 working days from the publication date to appear in PlumX Metrics, Semantic Scholar, and ResearchGate.
Abstract :
This paper presents a software-based Sign Language Interpreter aimed at improving communication between
hearing-impaired individuals and non-signers. The system uses MediaPipe for real-time hand landmark detection and
tracking, while a Convolutional Neural Network (CNN) model classifies gestures into corresponding text and speech outputs.
Focused on Indian Sign Language (ISL), the model is trained on a custom dataset to ensure accuracy under diverse lighting
and background conditions. By integrating deep learning with computer vision, the system achieves efficient and reliable
recognition without relying on additional hardware. This research highlights the potential of AI-based software applications
in fostering inclusivity and accessibility, offering an intelligent and cost-effective solution for assistive communication.
Keywords :
Sign Language Interpreter, MediaPipe, CNN, Deep Learning, Gesture Recognition, Assistive Communication.
References :
- G. Neve, and A. C. (2021). "Real-time Hand Gesture Recognition with MediaPipe and Deep Learning." Journal of Computer Vision and Pattern Recognition.
- A. Kapadia and M. Shah. (2020). "Deep Learning for Sign Language Recognition: A Survey." IEEE Transactions on Pattern Analysis and Machine Intelligence.
- L. Wei, et al. (2022). "A Novel Approach for Indian Sign Language (ISL) Recognition using Pose and Hand Tracking." Proceedings of the International Conference on Computer Vision (ICCV).
- J. Amin, M. Sharif (2021). "Sign Language Recognition using 3D Convolutional Neural Networks." Journal of Medical Imaging and Health Informatics.
- S. Kumar, A. Choudhary. (2021). "A Lightweight CNN for Sign Language Recognition on Mobile Devices." IEEE Xplore.
- C. Camgoz, O. Koller, et al. (2020). "Sign Language Transformers: A New Era in CSLR." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
- Google AI. (2020). "MediaPipe: A Framework for Building Perception Pipelines." Google AI Blog. https://ai.googleblog.com/2019/08/mediapipe-framework-for-building.html.
- O. Koller, et al. (2019). "Weakly Supervised Learning with Multi-Stream CNNs for Sign Language Recognition." Proceedings of CVPR.
- R. P. D. (2019). "WLASL: A Large-Scale Word-Level American Sign Language Video Dataset." Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops.
This paper presents a software-based Sign Language Interpreter aimed at improving communication between
hearing-impaired individuals and non-signers. The system uses MediaPipe for real-time hand landmark detection and
tracking, while a Convolutional Neural Network (CNN) model classifies gestures into corresponding text and speech outputs.
Focused on Indian Sign Language (ISL), the model is trained on a custom dataset to ensure accuracy under diverse lighting
and background conditions. By integrating deep learning with computer vision, the system achieves efficient and reliable
recognition without relying on additional hardware. This research highlights the potential of AI-based software applications
in fostering inclusivity and accessibility, offering an intelligent and cost-effective solution for assistive communication.
Keywords :
Sign Language Interpreter, MediaPipe, CNN, Deep Learning, Gesture Recognition, Assistive Communication.