Sie sind auf Seite 1von 4

Communication Aid for Deaf and Dumb

Guided by S.Indhumathi (Asst. Professor O.G)

Sabresh Mouli Mrinal Bage


ECE Department ECE Department
SRM Institute of Science and Technology, SRM Institute of Science and Technology,
Chennai, India-600026 Chennai, India-600026
sabresh.mouli@gmail.com mrinalbage123@gmail.com

Abstract— The Deaf and Dumb are facing many  To establish a two-way communication system using
challenges, and are craving for self-dependence. The above deep learning and computer vision.
mentioned cause was using flex sensors to record their  To train the system with 26 alphabet sign language
gestures. But we are taking this a step further and creating with different hand textures and to use the same
a platform in which Deaf and Dumb can interact with a database for the two-way communication.
normal person. They can use the sign language to  The use of artificial intelligence will make two-way
communicate with the normal person, and the sign communication possible for deaf and dumb.
language will be converted into text and the other person’s
text message will be converted into sign language. The
technology we have used is computer vision and deep
learning which comes under the broad umbrella of II. RELATED WORK
artificial intelligence. The processor which is going to be
To date many researchers have proposed several methods to
used is Raspberry Pi. The programming language that will
solve the problem of communication for deaf and dumb.
be used is Python. The technology will be useful in all the
These methods are discussed below in this section:
public sector place in which they can get what they want.
As the name suggests ,Wireless Bell for Deaf and Dumb. But
This will help the recruiters in recruiting them and a deaf
how can a deaf or a dumb respond to bells. A Deaf cannot
and dumb person can present papers in large seminars
hear to the bell and a dumb cannot respond to the calling bell.
and conferences.
Therefore, a simple low cost feasible system can be used to
Keywords—Artificial Intelligence, Computer Vision, help the deaf and the dumb to respond effectively to the
Deep Learning person knocking at the door. As the deaf cannot hear, a device
with vibrators is attached to their body in the form of some
I. INTRODUCTION wrist band or a watch. Similarly, for the dumb, a kind of bell
with speakers is developed which will playback some voice to
The sign language used as a means of communication between the person standing at the door as an acknowledgement
deaf and dumb is unadapted for private communications, as message. The entire system is wireless and can work
the language spoken to one is spoken to all present who effectively within a radius of 50m i.e. the range of RF
understand it. Spoken language can be whispered, or its Transmitter.
volume can be so reduced as to be inaudible to other ears than
those for which it is intended; but the force of the sign Traditionally, gesture recognition method was divided into
language cannot thus be modified, and when private two categories namely vision based and sensor based method.
conversations are held, written language is generally In vision based method, the computer camera is an input
employed. Besides the tediousness of this process cannot be device for varies gestures of hands and figures. In sensor
always resorted. based systems, gloves are used which can achieve the accurate
To date, several researchers have tried to tackle the problem of positions of hand gesture. Lots of studies have been done on
communication for deaf and dumb but haven’t been able to sensor-based approaches like gloves, helmets etc. But wearing
establish a two-way communication between them. Two-way it continuously is not possible. Therefore further work is
communication can be established with the help of computer concentrated on image-based approaches.
vision, deep learning and machine learning. The overall
contribution of this paper is as follows: The existing system suffers from lots of challenges in
providing two-way communication. Another problem is in the
accuracy of the existing systems. The existing system suffers
from accuracy problems due to occlusion of objects and due to
different textures of the hand, there is a deviation in the output
of the system.

III. HARDWARE REQUIREMENTS


A. Raspberry Pi:
Raspberry Pi is a credit-card sized computer that connects to a
computer monitor or TV and uses input devices like keyboard
and mouse. It is capable of performing various functionalities
such as surveillance system, military applications, surfing Fig 2: Logitech Camera
internet, playing high definition videos, live games and to
make data bases The paper is implemented using a Raspberry IV. PROJECT OVERVIEW
Pi 3 board and their specifications are as follows: We aim for developing the prototype model for deaf and dumb
1) Memory:1 GB LPDDR-900 SD-RAM (i.e. 900MHz) people by employing in a single compact device. The project
2) Processor: Raspberry Pi has a Broadcom BCM2837 system provides a unique solution for these people to manage their
on1chip module. It has quad core ARMv8 CPU. It can run at sites by themselves. The project is catered with the source
1.2 GHz video core for multimedia GPU. code of Python. It is the easiest programming language to
3) Power: The Pi requires a 5V power supply. interface with the Raspberry Pi. The project is run by the
4) GPIO (General Purpose Input-Output): GPIO is a generic source code of Python to assist dumb and deaf people in a
pin on an integrated circuit which can be configured as input single device which is so compact and easy for them to
and output pin. The Raspberry Pi board has 40 pins, 2.54mm manage.
expansion header.
5) SD card: Raspberry Pi has no storage on board so an
external memory is required to store the OS.
6) Connections: 4 USB ports,10/100 Mbps Ethernet and
802.11n wireless LAN connections,3.5mm jack for audio
out,BCM43438 Wi-Fi, Bluetooth low energy(BLE).

Fig 3: Text to Sign conversion output

The process started with the capturing the image and cropped
Fig 1: Raspberry Pi Version 3 the useful portion. Convert the RGB image into gray scale
image for better functioning, Blur the cropped image through
B. Logitech Camera Gaussian blur function and pass it to the threshold function to
It is a plug and play setup which is easy to apply. You can get the highlighted part of the image. Find the contours and an
easily make video calls on major IMs. It has a 5MP camera angle between two fingers. By using convex hull function we
with high resolution. It has built in mikes with right sounds can implement the finger point. Count the number of angles
gives you a clear conversation without any noise. XVGA which is less than 90 degree which gives the number of
video recording system has a reach of about 1024x768 defects.
resolution. In this project, we are using a Logitech camera Background Extraction is done through using histograms of
which is capturing the image and gesture control function. oriented gradient algorithm. The histogram of oriented
gradients (HOG) is a feature descriptor used in computer
vision and image processing for the purpose of object
detection.
Random Forests are an ensemble learning method for
classification, regression and other tasks, that operate by
constructing a multitude of decision trees at training time and
outputting the class that is the mode of classes or mean
prediction of the individual trees.

V. BACKGROUND EXTRACTION
Background Extraction is done through using histograms of
oriented gradient algorithm. The histogram of oriented
gradients (HOG) is a feature descriptor used in computer
vision and image processing for the purpose of object
detection. The technique counts occurrences of gradient
orientation in localized portions of an image. This method is
similar to that of edge orientation histograms, scale-invariant
feature transform descriptors, and shape contexts, but differs
in that it is computed on a dense grid of uniformly spaced cells
and uses overlapping local contrast normalization for Fig 5: Precision of the Random Forest Algorithm
improved accuracy. The essential thought behind the
histogram of oriented gradients descriptor is that local object The precision or accuracy is 0.77 i.e 77%. The benchmark
appearance and shape within an image can be described by the value is 70% but we have obtained 7% more accurate result.
distribution of intensity gradients or edge directions. The
image is divided into small connected regions called cells, and
for the pixels within each cell, a histogram of gradient VII. CONCLUSION
directions is compiled. We have designed the prototype model for deaf and dumb
people into a single compact device. The advantage of this
device is that it can be easily portable due to its less weight
and size. This paper fulfils the hand gesture recognition
process with some limitations as both the hands cannot be
used in this technique because the result is not efficient.

VIII. FUTURE SCOPE


The system can be further expanded for the alphabet, numbers
in gesture control. The input can be also taken in the form of
videos and the videos are divided into frames and the frames
can be converted into text. We can also add grammatically
structure for sign language. The system can be made handy by
incorporating it into a mobile phone.

IX. REFERENCES
Fig 4: Bounding Box
[1] Yiwen He, Jianyu Yang, Zhanpeng Shao, Youfu Li, ”Salient
VI. RANDOM FOREST ALGORITHM feature point selection for real time RGB-D hand gesture
Random Forests are an ensemble learning method for recognition”, IEE International Conference on real- time Computing
classification, regression and other tasks, that operate by and Robotics, 2017
constructing a multitude of decision trees at training time and [2] Rania A. Elsayed, Mohammed S. Sayed, Mahmoud I. Abdalla,
outputting the class that is the mode of classes or mean “Hand gesture recognition based on dimensionality reduction of
prediction of the individual trees. histogram of oriented gradients”, Japan-Africa Conference on
The training algorithm for random forest applies the general Electronics, Communication and Computers, 2017
technique of bootstrap aggregating, or bagging to tree learners. [3] Himadri Nath Saha, Sayan Tapadar, Shinjini Ray, Suhrid Krishna
Chatterjee, “A Machine Learning based approach for Hand Gesture
Given a training set X = x1, ..., xn with responses Y=y1,…,yn
Recognition using distinctive feature extraction”, IEEE 8th Annual
bagging repeatedly (B times) selects a random sample with Conference on Computing and Communication, 2018
replacements of the training sets and fits the trees to the
samples.
 

Das könnte Ihnen auch gefallen