Sie sind auf Seite 1von 6

International Conference on Computing and Control Engineering (ICCCE 2012), April, 2012 Institute of Electrical and Electronic Engineers

(IEEE), 1212 && 1313 April, 2012

Design and Implementation of Voice Control System for Wireless Home Automation Networks
1

Anjugam M, PG Scholar, 2Kavitha M, Assistant Professor, Veltech Dr.RR & Dr.SR Technical University, Chennai.
anjugam14@gmail.com, kavithamunisamygmail.com

ABSTRACT:
This paper presents the design and implementation of a voice control system for wireless home automation networks. Home automation industry is growing rapidly. This is fuelled by the need to provide supporting systems for the people, especially the elders, blinds and people with disabilities. The overall design of a wireless home automation system(WHAS) which has been built and implemented. In this systems one or more voice recognition modules have been added to the wireless networks. The automation centres on recognition of voice commands & uses zigbee communication modules. The voice command is a person independent. The voice sample is stored in database along with the function key. The voice received is processed and extract feature of the voice and find the matching sample from the database. The voice is processed and then the feature is extracted based on the MFCC algorithm. The matched samples generate the corresponding function key and transmit through zigbee module. The receiver is attached to the microcontroller. The microcontroller then processes the received data and switches the respective appliance.

INDEX TERMS: Home automation, speech recognition, feature extraction, MFCC algorithm. INTRODUCTION: Home automation designates an emerging practice of increased automation of household appliances and features in residential dwellings, particularly through

electronic means that allow for remote controlling. In these days, many household appliances such as computers, televisions, DVD players, home theatre systems, stereo components, air conditioners are popular. These devices are generally operated using remote controls based on IR (infra-red) signal .However, each remote/device usually uses a different device code, which prevents signals from being crossed. Although a universal remote control unit can transmit multiple device codes and control several home appliances in a house, its not necessarily optimal. Modern entertainment systems also have complicated control menus, which can require special buttons not found on the universal remotes. The elders, children or disabled people may be confused (find it laborious and inconvenient) when they need to find the remote controller and repeatedly press the control buttons to execute. In this paper, we propose a novel application of smart home automation .our work applies speech recognition technology based on MFCC algorithm to zigbee-based wireless networks.when speech recognition modules are deployed reasonably, users can give voice orders at any position of the house. For example, when you wake up in the morning ,even when you are still in bed, you say fan off, Open the window and light on .then the devices are automatically controlled. This system features its easy and flexible
1

International Conference Institute on Computing of Electrical and and Control Electronic Engineering Engineers (ICCCE (IEEE), 2012), 12 & 12 13 &April, 13 April, 2012 2012

operation, making it suitable especially for the elders and the disabilities, such as quadriplegia patients. SYSTEM OVERVIEW:

SPEECH RECOGNITION PROCESS . The speech recognition process includes two main stages. First stage contains feature extraction and storage of extracted features as training data. Second stage is test. In this stage, features of a new entered command are extracted. These features are used in order to make comparison with stored features to recognize command. MFCC algorithm is used for feature extraction and vector quantization method is used to reduce amount of achieved data in form of codebooks. These data are saved as acoustic vectors. In the matching stage, features of input command are compared with each codebook using Euclidean distance criterion.

figure 1.system overview The proposed voice control system architecture is shown in Figure 1. This system is realized on the basis of a wireless sensor and actuator network based on ZigBee protocol. The home automation network is composed of a (or several) voice recognition module(s), several sensor nodes, actuator nodes (such as wireless power outlet modules) and a network coordinator. Each node or module is added to the ZigBee network. The sensor nodes and the actuator nodes connected to each corresponding home appliance or even windows and curtains are deployed in the home environment and they autonomously form a mesh network. The network coordinator can provide multiple interfaces to the mesh network for local user interactions and remote network management. When the user speaks to the microphone, the module recognizes the words and sends control messages to the target ZigBee module through the network. The corresponding ZigBee module will operate the connected appliance via the connected driver circuit.

Figure 2.Speech recognition process MFCC Block Diagram :

- -

2012 Published by Coimbatore Institute of Information Technology

Institute of Electrical and Electronic Engineers (IEEE), 12 && 13 April, 2012 International Conference on Computing and Control Engineering (ICCCE 2012), 12 13 April, 2012

MFCC consists of seven computational steps. Each step has its function and mathematical approaches as discussed briefly in the following: Step 1: Preemphasis This step processes the passing of signal through a filter which emphasizes higher frequencies. This process will increase the energy of signal at higher frequency. Y (n) = X (n) - 0. 95 X (n - 1) -----(1) Lets consider a = 0.95, which make 95% of any one sample is presumed to originate from previous sample. Step 2: Framing The process of segmenting the speech samples obtained from analog to digital conversion (ADC) into a small frame with the length within the range of 20 to 40 msec. The voice signal is divided into frames of N samples. Adjacent frames are being separated by M (M<N).Typical values used are M = 100 and N= 256. Step 3: Hamming windowing Hamming window is used as window shape by considering the next block in feature extraction processing chain and integrates all the closest frequency lines. The Hamming window equation is given as: If the window is defined as W (n), 0 n N-1 where N = number of samples in each frame Y[n] = Output signal X (n) = input signal W (n) = Hamming window, then the result of windowing signal is shown below: Y (n)= X (n)* W (n ) ------(2)

Step 4: Fast Fourier Transform To convert each frame of N samples from time domain into frequency domain. The Fourier Transform is to convert the convolution of the glottal pulse U[n] and the vocal tract impulse response H[n] in the time domain. This statement supports the equation below: ( ) [ ( ) ( )] ( ) ( ) If X (w), H (w) and Y (w) are the Fourier Transform of X(t), H(t) and Y(t) respectively. Step 5: Mel Filter Bank Processing The frequencies range in FFT spectrum is very wide and voice signal does not follow the linear scale. The bank of filters according to Mel scale and then performed. Mel scale filter bank consists set of triangular filters that are used to compute a weighted sum of filter spectral components so that the output of process approximates to a Mel scale. Each filters magnitude frequency response is triangular in shape and equal to unity at the centre frequency and decrease linearly to zero at centre frequency of two adjacent filters. Then, each filter output is the sum of its filtered spectral components. After that the following equation is used to compute the Mel for given frequency f in HZ: ( ) [ ( ) ] Step 6: Discrete Cosine Transform This is the process to convert the log Mel spectrum into time domain using Discrete Cosine Transform (DCT). The result of the conversion is called Mel Frequency Cepstrum Coefficient. The set of coefficient is called acoustic vectors. Therefore, each input utterance is transformed into a sequence of acoustic vector.

w(n) = 0 . 54 - 0 . 46 cos

-----(3)

- -

International Conference Institute on Computing of Electrical andand Control Electronic Engineering Engineers (ICCCE (IEEE), 2012), 12 & 1213 & April, 13 April, 2012 2012

Step 7: Delta Energy and Delta Spectrum The voice signal and the frames changes, such as the slope of a formant at its transitions. Therefore, there is a need to add features related to the change in cepstral features over time . 13 delta or velocity features (12 cepstral features plus energy), and 39 features a double delta or acceleration feature are added. The energy in a frame for a signal x in a window from time sample t1 to time sample t2, is represented at the equation below: [ ] Each of the 13 delta features represents the change between frames in the equation 8 corresponding cepstral or energy feature, while each of the 39 double delta features represents the change between frames in the corresponding delta features. ( ) ( ) ( ) FEATURE MACTHING: Each feature vector in the sequence X is compared with all the stored code words in codebook, and the codeword with the minimum distance from the feature vectors is selected as proposed command For each codebook a distance measure is computed, and the command with the lowest distance is chosen. One way to define the distance measure is to use the Euclidean distances: (( ) )

EXPERIMENTAL RESULTS:

Fig shows all our nodes and modules in the voice control system. On this basis, a ZigBee-based Home Automation Network is established. Finally, all the functions described above are realized and tested. The proposed system is confirmed to be feasible.
The experiments are done as follow:
User: Light on. System: Success. User: Fan off. System: Success.

CONCLUSIONS: In this paper, a voice control system for wireless home automation networks is proposed. This system controls the multiple home appliances by using the speech recognition technology. This technology uses the MFCC algorithm for feature extraction and the matched features generate the corresponding function key and transmit through zigbee module. The receiver attached to the microcontroller and data received is taken as input data to the controller. The microcontroller then processes the received data and switches the respective appliance. This is fuelled by the need to provide supporting systems for the people, especially the elders, blinds and people with disabilities. References [1] Yusuf UZUNAY and Kemal BICAKCI,SHA: A Secure Voice Activated Smart Home for Quadriplegia Patients, IEEE International Conference
4

by finding the distance between the input vector X and each of the code words from the codebook C. The one with the smallest distance is coded as the output command. The matched command generate the corresponding function key and transmit through zigbee module. The receiver module attached to the microcontroller and data received is taken as input data to the controller. The microcontroller then processes the received data and switches the respective appliance.

Institute of Electrical Electronic Engineers (IEEE), 12 12 & 13 April, 2012 International Conference on Computing and and Control Engineering (ICCCE 2012), & 13 April, 2012

on Bioinformatics and Biomedicine, BIBM, 2007 [2] Guangming Song, Fei Ding, Weijuan Zhang and Aiguo Song,A Wireless Power Outlet System for Smart Homes, IEEE Transactions on Consumer Electronics, Vol. 54, No. 4, NOVEMBER 2008 [3] Wan-Ki Park, Intark Han, Kwang-Roh Park,ZigBee based Dynamic Control Scheme for Multiple Legacy IR Controllable Digital Consumer Devices, IEEE Transactions on Consumer Electronics, Vol. 53, No. 1, FEBRUARY 2007 [4] Il-Kyu HwangHome Network Configuring Scheme for All Electric Appliances Using ZigBee-based Integrated Remote Controller, IEEE Transactions on Consumer Electronics, Vol. 55, No. 3, AUGUST 2009 [5] Jinn-Kwei Guo, Chun-Lin Lu, Ju-Yun Chang, Yi-Jing Li,Ya-Chi Huang, Fu-Jiun Lu and Ching-Wen Hsu, Interactive Voice-Controller Applied to Home Automation, 2009 Fifth International Conference on Intelligent Information Hiding and Multimedia Signal Processing. [6] Tejaswini Hebalkar, Spring 2000 Voice Recognition and Identification System Final Report 18-551 Digital Communications and Signal Processing Systems Design [7] Nilsson Magnus, October 2001, Speaker Verification in JAVA, A thesis submitted in partial fulfillment of the requirements for the degree of Master of Computer and Information Engineering, School of Microelectronic Engineering, GriffithUniversity. [8] Jamal Price, sophomore student, Design an automatic speech recognition system using maltab, University of Maryland Estern Shore Princess Anne. [9] Ahmad Kamarul,Ariff Bin Ibrahim, Biomedical engineering laboratory student pack,UTM Jjohor [10] E.C. Gordon,Signal and Linear System Analysis.John Wiley &Sons Ltd., New York, USA,1998.

[11] Stan Salvador and Pjilip Chan,FastDTW: Toward Accurate Dynamic Time Warping in Linear time space,Florida Institute of Technology,Melbourne. [12] Chunsheng Fang, From Dynamic time warping (DTW) to Hidden Markov Model (HMM), University of Cincinnati,2009.

International Conference on Computing and Control Engineering (ICCCE 2012), 12 & 13 April, 2012

-1-

Das könnte Ihnen auch gefallen