Beruflich Dokumente
Kultur Dokumente
based on Biosignals
Yuji Takano
Kenji Suzuki
University of Tsukuba
1-1-1 Tennodai
Tsukuba, Japan
University of Tsukuba/JST
1-1-1 Tennodai
Tsukuba, Japan
yuji@ai.iit.tsukuba.ac.jp
kenji@ieee.org
ABSTRACT
We propose a novel wearable interface for sharing facial expressions between children with autism spectrum disorders
(ASD) and their parents, therapists, and caregivers. The developed interface is capable of recognizing facial expressions
based on physiological signal patterns taken from facial bioelectrical signals and displaying the results in real time. The
physiological signals are measured from the forehead and
both sides of the head. We veried that the proposed classication method is robust against facial movements, blinking,
and the head posture. This compact interface can support
the perception of facial expressions between children with
ASD and others to help improve their communication.
General Terms
Measurement
Keywords
Facial expression, Smile sharing, Autism Spectrum Disorder
1. INTRODUCTION
In this paper, we propose a novel interaction method for
sharing childrens facial expressions with their parents in
order to facilitate communication. In human communication, facial expressions carry some of the most important
non-verbal information. Facial expressions include psychological information, such as emotions, which are very important aspects of communication. People can read a persons
thoughts simply by observing their expression. Psychological studies have found that facial expressions can project
emotions such as disgust, sadness, happiness, fear, anger,
and surprise [2]. Expressing these emotions is a universal
Permission to make digital or hard copies of all or part of this work for personal or
classroom use is granted without fee provided that copies are not made or distributed
for prot or commercial advantage and that copies bear this notice and the full citation on the rst page. Copyrights for components of this work owned by others than
ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specic permission
and/or a fee. Request permissions from permissions@acm.org.
IDC14, June 1720, 2014, Aarhus, Denmark.
Copyright 2014 ACM 978-1-4503-2272-0/14/06 ...$15.00.
http://dx.doi.org/10.1145/2593968.2610455.
213
Socet
Electrode
Headband
Socket
Electrode
Looking
Vibration
LED
LED Display
Electrode
Vibrator
Smile sharing:
proposed interaction
markers requires preparation, which makes this approach laborious, and the markers can be easily occluded. Thus, the
development of a method for capturing facial expressions
that is easy to use and does not depend on spatial orientation is still a dicult challenge. We have been developing
a tool to detect the facial expressions of a person who has
diculty with expressing their intent in an accurate and continuous manner through the use of a wearable device. This
allows users to not only capture the facial expression but
also share them with others, even if the face is not always
observable by sensors installed in the environment, such as
cameras or depth sensors. In this paper, we propose the concept of smile sharing, where a wearable devicenamely, an
aective communication aidis used that meets the above
criteria to communicate facial expressions. We evaluated the
device to verify its performance through several case studies.
2.2
Smile Sharing
2. METHODOLOGY
The proposed system provides a novel method of interaction, particularly between children and their parents, that
considers use in daily life. Figure 1 shows a conceptual diagram of the proposed interaction frame. We rst describe
the method for capturing and classifying the facial expressions independent of spatial orientation and then the sharing
of the facial expressions.
3.
SYSTEM CONFIGURATION
The system consists of an interface unit and signal processing unit. The interface unit measures signals and outputs the classication results. The signal processing unit
classies facial expressions based on the measured sEMG
signals and sends the result to the interface unit via Bluetooth wireless communication.
3.1
Interface unit
We developed two dierent wearable interfaces: headmounted and wrist-mounted devices. Figure 2 shows an
overview of the head-mounted interface. The wrist-mounted
interface is a simple vibration device that simply vibrates
when the smile is detected by the head-mounted interface.
214
4.
Calibration
Class
Labeling
Input
A/D
fs=1000[Hz]
Training
Output
=100[ms]
Filtering
400
Rectification
and
ICA
Comb Band
b(t) Smoothing c(t)
Pass x(t)
4.1
y(t)
Signal Processing
Section
In order to evaluate the classication accuracy of the proposed system, we compared it to the human cognitive ability to recognize smiling in an experimental setting. We
recorded videos of three people (persons AC); each alternately smiled and had a neutral expression for two or three
times over about 20 s while wearing the headband interface.
Nine subjects (eight male, one female) in their twenties and
thirties were recruited for the experiment. Informed consent was obtained from the participants in advance. Videos
of the smiling/neutral faces (AC) were shown to the subjects, and they were asked to mark the smile intervals by
clicking a button to indicate the start and stop of smiles.
We covered the LED in the videos to avoid inuencing the
subjects judgment. We calculated the maximum, minimum,
and median values of precision and recall based on the classications by the subjects and proposed system. Figure 5
shows the results.
As shown in the gure, the precision of each subject was
above 0.95, but the recall varied among subjects depending
on who created the facial expressions. The dierences in recall may have been due to the dierent facial features, some
of which are more dicult to recognize than others. This
made it more dicult to set a threshold for smiling (as for
subject B), which lowered the recall. However, in terms of
classication accuracy, the results were positive because the
precision average was sucient for potential applications.
We also calculated the intra-class correlation coecient to
evaluate the degree of coincidence between the classication
by the interface and the judgment of the subjects. The average intra-class correlation coecient was more than 0.936,
which is also sucient.
LED
Bluetooth
Bluetooth
Measuring
Section
SVM
Feature
Vector
Electrode
EXPERIMENT
We conducted two experiments to evaluate the performance of the proposed system. In this section, we present
the classication accuracy and robustness against head motion of the system.
Light Emitting
Section
4.2
Evaluation of robustness
We then conducted an experiment to investigate the robustness of the system against head motion artifacts. In
this experiment, we investigated whether the system is capable of classifying facial expressions when there are disturbances such as head motions. The classication accuracy
against head nodding (forward and back head movement),
head tilting (left and right head tilting), head shaking (right
and left rotation), and blinking was checked to evaluate the
robustness of the system. The three motions we investigated
(nod, tilt, and shake) correspond to all possible motions
215
Preci
si
on/Recal
l
the acquired signals from the sides of the head and forehead
can be used for facial expression classication. Through several experiments, we veried the classication accuracy of
the developed system. The results demonstrated that the
interface can be used in real environments with some disturbances to classify facial expressions with high accuracy
and to present smiles in real-time. Further investigation will
include the implementation of adaptive ltering to remove
motion artifacts.
So far, we have presented the concept of a novel interaction design between children and their parents and developed interfaces that enable the realization of such interaction. We have already conducted a feasibility study with
children having ASD during robot-assisted activities and
conrmed that the proposed device is acceptable [7]. In the
future, we plan to conduct a user study with children and
families to verify that the interfaces can support the sharing
and perception of facial expressions in the given scenario.
n=9
0.9
0.8
Preci
si
on
Recal
l
0.7
Accuracy[%]
100
90
6.
80
neutral
smile
n=8
70
no motion
blink
nod
tilt
REFERENCES
shake
216