Authors :
Lisha Kurian; Anaj Pravin; Calvin Johnson; Abhishek Unnikrishnan; Aswin Sunil
Volume/Issue :
Volume 9 - 2024, Issue 11 - November
Google Scholar :
https://tinyurl.com/4byvdj65
Scribd :
https://tinyurl.com/29a6m6tr
DOI :
https://doi.org/10.38124/ijisrt/IJISRT24NOV518
Note : A published paper may take 4-5 working days from the publication date to appear in PlumX Metrics, Semantic Scholar, and ResearchGate.
Abstract :
The project is to enable people who are not
versedin sign language or people from the deaf or hard-
of-hearing community to communicate by using a system
that translates their American Sign Language (ASL)
gestures into text, which could then be converted into
speech. Computer vision and machine learning
algorithms allow the system to “read” the sign language
as accurately as possible, and then translate into a native
text. Text is transcribed to speech using Text-to-Speech
(TTS) capabilities The proposed calibration can be
applied to real-time applications serving purpose for
accessible and decent spoken communication among
different individuals with hearing loss which applies the
natural co-articulation constraints in various social or
professional environments.
Keywords :
Component, Formatting, Style, Styling, Insert.
References :
- Truong, V.N., Yang, C.K. and Tran, Q.V., 2016, October. A translator for American sign language to text and speech. In 2016 IEEE 5th Global Conference on Consumer Electronics (pp. 1-2). IEEE.
- Camgoz, N.C., Koller, O., Hadfield, S. and Bowden, R., 2020. Sign language transformers:Joint end-to-end sign language recognition and translation. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (pp. 10023-10033).
- Amrutha, K. and Prabu, P., 2021, February. ML based sign language recognition system.In 2021 International Conference on Innovative Trends in Information Technology (ICITIIT) (pp. 1-6). IEEE.
- Chen, M., Tan, X., Li, B., Liu, Y., Qin, T., Zhao, S. and Liu, T.Y., 2021. Adaspeech: Adaptive text to speech for custom voice. arXiv preprint arXiv:2103.00993.
- Zhou, Z., Chen, K., Li, X., Zhang, S., Wu, Y., Zhou, Y., Meng, K., Sun, C., He, Q., Fan, W. and Fan, E., 2020. Sign-to-speech translation using machine-learning-assisted stretchable sensor arrays. Nature Electronics, 3(9), pp.571-578.
- Cheng, K.L., Yang, Z., Chen, Q. and Tai, Y.W., 2020. Fully convolu- tional networks for continuous sign language recognition. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XXIV 16 (pp. 697-714).Springer Inter- national Publishing.
- Tan, X., Chen, J., Liu, H., Cong, J., Zhang, C., Liu, Y., Wang, X., Leng, Y., Yi, Y., He, L. and Zhao, S., 2024. Naturalspeech: End-to-end text-to-speech synthesis with human-level quality. IEEE Transactions on Pattern Analysis and Machine Intelligence.
- Tiku, K., Maloo, J., Ramesh, A. and Indra, R., 2020, July. Real-time conversion of sign language to text and speech. In 2020 Second International Conference on Inventive Research in Computing Applications (ICIRCA) (pp. 346-351). IEEE.
- Kim, T.H., Cho, S., Choi, S., Park, S. and Lee, S.Y., 2020, May. Emotional voice conversion using multitask learning with text-to-speech. In ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) (pp. 7774-7778). IEEE.
- Kahlon, N.K. and Singh, W., 2023. Machine translation from text to sign language: a systematic review. Universal Access in the Information Society, 22(1), pp.1-35.
- Ren, Y., Hu, C., Tan, X., Qin, T., Zhao, S., Zhao, Z. and Liu, T.Y., 2020. Fastspeech 2: Fast and high-quality end-to-end text to speech. arXiv preprint arXiv:2006.04558.
- Makhmudov, F., Mukhiddinov, M., Abdusalomov, A., Avazov, K., Khamdamov, U. and Cho, Y.I., 2020. Improvement of the end-to-end scene text recognition method for “text-to-speech” conversion. International Journal of Wavelets, Multiresolution and Information Processing, 18(06), p.2050052.
The project is to enable people who are not
versedin sign language or people from the deaf or hard-
of-hearing community to communicate by using a system
that translates their American Sign Language (ASL)
gestures into text, which could then be converted into
speech. Computer vision and machine learning
algorithms allow the system to “read” the sign language
as accurately as possible, and then translate into a native
text. Text is transcribed to speech using Text-to-Speech
(TTS) capabilities The proposed calibration can be
applied to real-time applications serving purpose for
accessible and decent spoken communication among
different individuals with hearing loss which applies the
natural co-articulation constraints in various social or
professional environments.
Keywords :
Component, Formatting, Style, Styling, Insert.