Beruflich Dokumente
Kultur Dokumente
ABSTRACT
Mankind has grown from Keyboard input to touch. Now the big question is, What Next? In this paper a solution
to this problem is proposed by introducing a new algorithm so as to minimize the hassle of controlling the most frequently
used set of computer applications. The proposed algorithm provides the flexibility to control different application
depending upon hand gesture. A user can assign his/her desired application to the hand gesture and can control the
application just by posing the assigned gesture in front of the computer. Since, the proposed approach depends only on the
peaks generated due to fingers raised hence, it is invariant to users skin tone as well as the lighting condition in the
background. A simple web cam working on 30 fps with up to 30 mega pixel intensity is sufficient to perform the task.
The images are captured through web cam. Further, pre-processing is performed for background noise removal and
segmenting the hand object from background. Likewise, only segmented hand object is extracted and is processed for
shape based features computation. As the speed dial application of smart phones minimizes the time and effort to open
contact detail, this paper work on similar lines.
KEYWORDS: Image Processing, Hand Gesture Recognition, Human Computer Interaction, Controlling Applications,
Centroid, K Curvature, Computer Vision
INTRODUCTION
Motivation: Computer applications are controlled with the help of keyboard, mouse, joystick etc. and nowadays
it is controlled our with a touch. Hand gestures are the next logical step towards making the control easier. Gestures are
motion of the body/face/hand or any physical action form by the user in order to convey some information to someone else.
Gesture recognition is the process by which gesture made by the user is made known to the system, through the use of
computer vision via webcam connected to the computer. There is great emphasis on using hand gesture as a substitute of
new input modality in broad range applications. Hand gestures have the natural ability to represent ideas and actions very
easily, thus using these different hand shapes, being identified by gesture recognition system and interpreted to generate
corresponding event, has the potential to provide a more natural interface to the computer system. This type of natural
interaction is the core of immersive virtual environments. It can be simply realize that a wide range of gestures are utilized
in our daily personal communication. People gesticulate more when they are talking on their cell phones and are not able to
see each other as in face to face communication. In fact, we gesticulate twice more than we speak. The gestures vary
greatly among cultures and context still are intimately used in communication. The significant use of gestures in our daily
life as a mode of interaction motivates the use of gestural interface and employs them in wide range of application through
computer vision.
www.tjprc.org
editor@tjprc.org
38
Related Work: There are many gesture recognition techniques developed for tracking and recognizing various
hand gestures. Each one of them having their own advantages and disadvantages. First of them is wired technology,
in which users need to tie up themselves with the help of wire in order to connect or interface with the computer system.
In wired technology user isnt able to move freely as they are connected with the computer system via wire. The issue was
the length of the wire. Like a telephone it does not give us the freedom and therefore have limited movement. Then there
are Instrumented gloves or electronics gloves or data gloves. These instrumented gloves are made up of some sensors,
which provide the information related to hand location, finger position etc. through the use of sensors. The advantage of
using sensors is that it almost never goes wrong. These data gloves provide good results but they are extremely expensive
to utilize in wide range of common application. Data gloves are then replaced by optical markers. These optical markers
project Infra-Red light and reflect this light on screen to provide the information about the location of hand or tips of
fingers wherever the markers are wear on hand, the corresponding portion will display on the screen. These systems also
provide the good result but require very complex configuration. Later on some advanced techniques have been introduced
like Image based techniques which requires processing of image features like texture, color etc. If we work with these
features of the image for hand gesture recognition the result may vary and could be different as skin tones and texture
changes very rapidly from person to person from one place to another in the world. And also under different illumination
condition, color texture gets modified which leads to changes in observed results. For utilizing various hand gesture to
promote real time application this paper works on vision based hand gesture recognition system that work on shape based
features for hand gesture recognition. Every person poses almost similar hand shape with one thumb and four fingers
under. The proposed approach is designed and implemented for working on hand gestures.
Figure 1
Impact Factor (JCC): 4.9467
39
Pre Processing
Image Pre-processing is necessary for image enhancement. Our cameras take images in RGB format. In this
algorithm, the input RGB image gets converted in to YCbCr images as RGB color space is more sensitive to different light
conditions YCbCr is not an absolute color space; rather, it is a way of encoding RGB information. The actual color
displayed depends on the actual RGB primaries used to display the signal. Therefore a value expressed as YCbCr is
predictable only if standard RGB primary chromaticity is used. Cathode ray tube displays are driven by red, green, and
blue voltage signals, but these RGB signals are not efficient as a representation for storage and transmission, since they
have a lot of redundancy. Redundancy is the number of bits used to transmit a message minus the number of bits of actual
information in the message. It is the amount of wasted space used to transmit certain data [2].
Where Y is the luma component and CB and CR are Chroma components. Here, the prime symbols mean gamma
correction is being used; thus R, G and B nominally range from 0 to 1, with 0 representing the minimum intensity
(for black) and 1 the maximum (for white). The following figure shows RGB and YCbCr images.
B) Ac
B is the symmetric structuring element. The algorithm terminates at the iteration step k if Xk= Xk-1. The set Xk
then contains all the filled holes; the union of Xk and A contains all the filled holes and their boundaries contains all the
filled holes. After removing noise it uses a fuzzy filter. Fuzzy filters provide promising result in image-processing tasks
that cope with some drawbacks of classical filters. Fuzzy filter is capable of dealing with vague and uncertain information.
Image Taken
editor@tjprc.org
40
Then convert the noiseless image to binary image. In binary images, each pixel is stored as a single bit 0 or 1.
Figure 3 shows the binary images.
41
Where Mijis the image moment x and y are the coordinates of the pixels I (x,y) is the intensity of images pixel at
(x,y) order of the moment is i+j
Centroid, C (a,b) = (M10/M00,M01/M00)
Where M10=Sum x M01=Sum y M00=area of the binary image (a,b) are the coordinates of the centroid
Figure 7: Centroid
Peaks Detection: Peaks are found by finding the tips of the fingers to find the number of fingers raised. Boundary
matrices of the segmented binary image are obtained in the pre-processing part. This part is performed over the hand
except the thumb region that is the binary
bina hand image is localized such that the part under consideration is only the part
except the thumb region. Here it calculates the local maximas and the local minimas. For a vertically raised hand,
y coordinates of the binary image are traced,
trace that is, the rows of the binary matrix and for a horizontally raised hand the x
www.tjprc.org
editor@tjprc.org
42
coordinates of the binary image are traced. To find local maximas and minimas in both the cases. For vertically raised
hand, we take the average of all the local minimas of y coordinates,
where m is the y coordinate of the reference point, R. for x coordinate the midpoint of the localized hand is taken,
that is, (m,n) gives us the reference point R if the hand object is raised vertically.
(m, n) gives us the reference point R if the hand object is raised vertically.
For horizontally raised hand, the average of the local minimas of x coordinates is taken,
where m is the x coordinate of the reference point, R. for y coordinate we take the mid point,
(m,n) gives the reference point R if the hand object is raised horizontally.
Figure 9: Length RC
Also the Euclidean distance between Centroid C (a, b) and each local maximas is calculated, that is, every peak.
Say Q (x,y) is a local maxima then,
43
www.tjprc.org
editor@tjprc.org
44
00000
93
00001
91
00010
89
00011
91
00100
92
00101
88
00110
90
00111
91
01000
91
01001
91
01010
92
01011
89
45
Table 1: Contd.,
01100
91
01101
92
01110
89
01111
90
10000
91
10001
92
10010
91
10011
90
10100
91
10101
92
10110
93
10111
92
11000
91
11001
95
11010
96
11011
92
11100
94
11101
90
11110
89
11111
95
CONCLUSIONS
We proposed the shape based approach for hand gesture recognition to run multiple applications with just
movement of our hand. This paper first pre-processed the image and then finds centroid and the reference point. Lengths
between reference point, centroid and peaks are calculated, if the length of peaks is found greater then it is a peak. Number
of peaks detected gives the number of fingers raised and it is checked if the thumb is also raised or not, which is assigned
www.tjprc.org
editor@tjprc.org
46
to different applications that can be opened by the user. These applications are user defined. It works just as speed dial in
our phones. This can save a lot of time and works perfectly for people as time is of the essence. We can increase the
number of application that we can open with just these gestures by taking the values as binary and assigned these values to
the variable. The proposed algorithm is simple and independent of user characteristics. In future, work on both the hands
can be done.
REFERENCES
1.
Hand Segmentation Techniques to Hand Gesture Recognition for Natural Human Computer Interaction published
in International Journal of Human Computer Interaction(IJHCI), Volume (3) : Issue (1) : 2012
2.
Hand Gesture Recognition based on Shape Parameters Meenakshi Panwar Centre for Development of Advanced
Computing Noida, Uttar Pradesh, India
3.
Meenakshi Panwar and Pawan Singh Mehra, Hand Gesture Recognition for Human Computer Interaction,
in Proceedings of IEEE International Conference on Image Information Processing(ICIIP 2011), Waknaghat,
India, November 2011.
4.
Rajeshree Rokade, Dharmpal Doye, Manesh Kokare, Hand Gesture Recognition by Thinning Method,
in Proceedings of IEEE International Conference on Digital Image Processing (ICDIP), Nanded India, pages
284 287, March 2009.
5.
Computer Music Controller Based on Hand Gestures Recognition Through Web-cam Junhao Jiang, Junji Ma,
Yiye Jin Department of Electrical Engineering, Stanford University