Real-Time Gesture Recognition for Sign Language Using Multimodal Techniques
DOI:
https://doi.org/10.47392/IRJAEH.2025.0238Keywords:
Accessibility, Multimodal Learning, Mediapipe, Continuous Sign Language, Real-Time Gesture Detection, Sign Language RecognitionAbstract
Sign language serves as a crucial communication medium for individuals with hearing and speech impairments. However, existing recognition systems primarily focus on isolated word detection, limiting their effectiveness for real-time communication. This paper presents a Continuous Sign Language Recognition System using a multimodal approach to seamlessly interpret sign language gestures into coherent words. The system leverages Mediapipe, a robust computer vision framework, to detect and track hand gestures, body posture, and facial landmarks in real time. By processing keypoints extracted from these modalities, the system enhances recognition accuracy and ensures fluid word construction. Unlike traditional methods that rely on gloves or additional hardware, this approach offers a cost-effective and accessible solution for real-time sign language translation. Experimental evaluations demonstrate the system’s ability to accurately recognize continuous gestures, addressing challenges such as real-time performance, word-level interpretation, and scalability. The proposed model significantly improves accessibility and inclusivity, bridging the communication gap between sign language users and non-signers.
Downloads
Downloads
Published
Issue
Section
License
Copyright (c) 2025 International Research Journal on Advanced Engineering Hub (IRJAEH)

This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.