I-SRAVIA-Breaking the Wall of Silence: Prototyping Testing and Validating.

Uloženo v:
Podrobná bibliografie
Název: I-SRAVIA-Breaking the Wall of Silence: Prototyping Testing and Validating.
Autoři: Mishra, Shraavya, Mishra, Sumona Karjee
Zdroj: IEOM India Conference Proceedings; 11/6/2025, p151-160, 10p
Témata: SIGN language, COMPUTER vision, HUMAN-computer interaction, COMMUNICATION barriers, VISUAL communication, ASSISTIVE technology, MACHINE learning
Geografický termín: INDIA
Abstrakt: Approximately 63 million individuals in India have significant hearing and speech impairments. This creates substantial communication barriers that restrict social, educational, and occupational inclusion. Current solutions rarely support bidirectional, real-time interaction tailored for Indian Sign Language (ISL). In this work, we introduce I-SRAVIA (Indian Sign-Language Responsive and Voice Intelligent Assistant), a computer vision-driven prototype enabling two-way communication between ISL users and hearing individuals.The system employs a dataset of 1,149 ISL gesture images across nine classes captured by a webcam. The image dataset was augmented to 14,937 samples by treating the real dataset with various parameters. The feed-forward Multi-Layer Perceptron (MLP) sequential model specialized for Convolutional Neural Network (CNN) to classify images into nine ISL words for the model generation that. was trained over 25 epochs (batch size 32). Software evaluations-comprising training/validation accuracy, loss metrics, and confusion matrix analysis-demonstrated near-perfect performance with zero misclassification among five tested gestures. Real-time trials involving 160 gesture inputs produced a Mean Magnitude of Relative Error (MMRE) of 15.6%, equivalent to 84.4% prediction accuracy. Orientation robustness tests confirmed reliable gesture recognition within ±25° deviations. The user interface was developed using Flask, HTML/CSS, and JavaScript using principles of human factors engineering. The interface supports both gesture-to-text/voice and voice-to-text modes. These findings demonstrate the feasibility of a reliable, two-way ISL-based communication platform. Future work would expand the gesture lexicon library, leverage enhanced computational resources, and conduct usability testing in real-world environments to take this work toward real case deployment. [ABSTRACT FROM AUTHOR]
Copyright of IEOM India Conference Proceedings is the property of IEOM Society International and its content may not be copied or emailed to multiple sites without the copyright holder's express written permission. Additionally, content may not be used with any artificial intelligence tools or machine learning technologies. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)
Databáze: Complementary Index
Buďte první, kdo okomentuje tento záznam!
Nejprve se musíte přihlásit.