Python-Based Real-Time Sign Language Interpreter Using Computer Vision and Machine Learning
DOI:
https://doi.org/10.47392/IRJAEH.2025.0295Keywords:
Gesture, Haar Cascade, ClassifierAbstract
Humans communicate with one another using body language (gestures), such as hand and head gestures, facial expressions, lip movements, and so forth, or through natural language channels like words and writing. Sign language comprehension is just as crucial as knowing normal language. The primary means of communication for those who are hard of hearing is sign language. Without a translation, speaking with other hearing people can be difficult for those with hearing impairments. Because of this, the social lives of deaf people would be greatly improved by the installation of a system that recognizes sign language. In order to recognize the features of the hand in pictures captured by a webcam, we have presented in this study a marker-free, visual American Sign Language recognition system that makes use of image processing, computer vision, and neural network techniques. This paper deals with full phrase gestures that are used regularly every day and methods used to converted them to text. A number of image processing techniques have been used to identify the hand shape from continuous pictures. The Haar Cascade Classifier is used to determine the interpretation of signs and their associated meaning.
Downloads
Downloads
Published
Issue
Section
License
Copyright (c) 2025 International Research Journal on Advanced Engineering Hub (IRJAEH)

This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.