Sign Language Recognition Using Machine Learning A Survey
Sign Language Recognition Using Machine Learning A Survey
ISSN No:-2456-2165
Abstract:- It is said that more than one billion people in have sensors on them, and thus the movements are captured.
the world are disabled. One of the only ways they can Even though this system provides high accuracy it is not cost-
communicate among themselves or with people who don’t effective. Vision-based systems are the ones in which the
have this disability is sign language. Sign language is a signs are detected using visions, for this purpose either static
creative way to communicate within the deaf community images are used or real-time images are captured. Sign
by using gestures done by hand and other means that do language is all about perspective hence the vision-based is an
not require talking. It involves combining hand effective method.
movements, their shapes, etc. Body movements and facial
expressions are also taken into consideration. All these The proposed algorithm in this solution is CNN,
factors help convey the person’s thoughts in a fluent convolutional neural network that is a part of deep learning. It
manner. Most of the general public who aren’t disabled is mostly used for analyzing visual representations. Tensor
have no knowledge about sign language. Even out of the flow is also used which is a tool used to develop and train the
few who are aware, the majority don’t know how to use it model.
for communication, this stops them from interacting with
the deaf and mute people. Through this software, we want II. LITERATURE SURVERY(ABSTRACT)
to raise awareness towards sign language and help bridge [1]In this publication, an effort has been made to
the gap by creating a sign language interpreter that highlight the work of American Sign Language researchers
recognizes hand gestures. and compare their work. Asl - American sign language is
Keywords:- Sign Language Recognition, Mobile Application, used , as it is the most widely used sign language. This system
Convolutional Neural Network, Machine Learning. uses Microsoft Kinect for image acquisition or extracting the
dataset. Feature extraction is done using PCANet (Principal
I. INTRODUCTION component analysis network) and the classification of of ASL
letter is done using COnvolution neural network(CNN). The
Sign language is a way of communication for the deaf majority of SL and gesture identification problems have been
and sign language recognition is the method through which solved using statistical modeling techniques such as PCA and
the hand gestures in sign language are recognized by using support vector machines. The system proves to be 98%
images or videos. effective due to the fact that the application requires more
features or distinction to be accurate. Less number of signs
Each country has its own sign language. There are more were used to train the model, leading to poor user experience.
than 300 sign languages in the world, each differing from
another based on what part of the world they belong to. Sign [2]This system is a real-time sign language translator
language may even be different in the same country based on that translates sign language to text. It uses GoogLeNet
the accent. region,etc. One of the most prominent sign architecture to train the data .It uses Convolution neural
language systems known is ASL which is the American network to classify each frame in the video to a letter, and it
standard language. ASL is a natural language with its own reconstructs to display the most likely word from the
structure and unlike spoken English. It is a visual language classification.This is a a robust model that correctly classifies
that uses the movement and placement of hands to convey the letters a-e in the majority of cases when used by first-time
meaning of words. Facial expressions and body movements users, and another that correctly classifies letters a-k in the
also play an integral part in the language. majority of situations Given the dataset restrictions and the
promising findings obtained, we are hopeful that with further
In India ISL, Indian Sign Language is used. This is research and data, we will be able to build a fully
different from ASL as it uses its own vocabulary and generalizable translator for all ASL letters. Lack of variety in
grammar. Even though ISL is used in India, a part of the deaf the datasets, the validation accuracies were not directly
community still use ASL so for this proposed system we have replicable when tested on the web application. Better letter
decided to use ASL. segmentation is required, as well as a more smooth method for
retrieving photographs from users at a faster rate.
There are two different types of approaches to recognize
sign language, they are contact-based systems and vision- [3]Sign Language Recognition is one of today's fastest-
based systems. Contact-based systems as the name suggests growing disciplines of study. In these disciplines, many
are based on contact. Gloves are used in these systems, which innovative techniques have lately been developed. For the
The below list (TABLE I) outlines survey of papers related to the topic in brief with possible gaps/limitations within the
proposed system.
Static sign language Lean Karlo S. A sign language learning This system can
recognition using Tolentino, Ronnie system which is based on recognize only static
deep learning O. Serfa Juan, the modelling technique of sign languages.
[1] August C. Thio-ac, 2019 skin color. A specific range
Maria Abigail B. is agreed upon that will
Pamahoy, Joni extract pixels from non
Rose R. Forteza, pixels.
and Xavier Jet O.
Garcia
IV. EXISTING SOLUTION processing since sign language comprises both static and
dynamic gestures.
ASL recognition isn't a new problem in computer vision.
Researchers have utilized classifiers from a range of Another existing solution for communication for hearing
categories over the last two decades, which we may generally impaired people is a chat application. Chat programmes have
divide into linear classifiers, neural networks, and Bayesian evolved into a great tool for individuals to connect with one
networks. The research was also based on a variety of input another in a variety of languages. There are many different
sensors, gesture segmentation, feature extraction, and chat software that are utilized by different individuals in
classification approaches. different languages, but there isn't one that allows you to
interact with sign languages. The recent release of low-cost
The proposed system analyzes the American Sign depth sensors, such as the widely used Microsoft Kinect
Language gestures and then converts them into human sensor, has facilitated the development of new gesture
readable text.The majority of the works designed to address detection algorithms. Depth photographs generate a three-
this problem have used one of two approaches: contact-based dimensional model of the scene, which can be used to ease
systems, such as system gloves, or vision-based systems, tasks like people segmentation and tracking, body component
which rely solely on cameras. Contact based method is the detection, and motion prediction.
one where the signer has to wear a hardware glove and do the
signs , and the hand movements get captured . This system is V. PROPOSED SOLUTION
uncomfortable for practical and daily use, despite having an
accuracy of 90%.Static and dynamic recognition are two types Developing an android application aided by machine
of vision-based methods. Static is concerned with the learning techniques to recognize hand gestures can be done
identification of static gestures (two-dimensional images), with ease. Hand gesture recognition shows how fast the
whereas dynamic is concerned with the capturing of motions algorithms detect the gestures in a single shot. The faster and
in real time. This entails the employment of a camera to stable it can be, the user experience will be smoother and
record motion.Vision is a key factor in sign language, and better. The proposed SLR system constitutes the first attempt
every sign language is intended to be understood by one to merge a vision-based approach (i.e., processing of images)
person located in front of another, from this perspective, a with the accurate extraction of skeletal data without
gesture can be completely observable. employing data gloves or other sensors that limit the
Sensor-based devices, such as SignSpeak, were movements of a signer. The application opens to a camera,
employed in many research. This device used a variety of which detects the hand movements of the signer. Hand gesture
sensors, including flex and contact sensors for finger and palm features in the uploaded images are extracted and used to
movements, as well as accelerometers and gyros for hand recognize the type of gesture.
movement; the gloves were then trained to recognise different
gestures using Principal Component Analysis, and each
gesture was then classified into alphabets in real time.In
visual-based SLR, several strategies have been
established.Many people experimented with image and video
REFERENCES