Beruflich Dokumente
Kultur Dokumente
1.1
1.2
A third effect often mentioned in connection with object recognition across different viewpoints relates to what is known as the
canonical view. A canonical view of an object can be named better, faster and more accurately than other views of the same object
(Palmer, Rosch, & Chase, 1981). In face perception research, it
has long been speculated as to whether canonical views also exist for faces similar to those found for objects. A candidate for a
possible canonical view is the 3/4 view (e.g. OToole, Edelman,
& Bulthoff, 1998), located approximately midway between a fullface (or frontal) view and a profile view. However, there is no clear
evidence showing that one specific canonical view exists for face
20
constituent parts (such as eyes, nose, mouth, etc.) is used to isolate local part-based information. Previous studies have often attempted to directly alter the facial features or their spatial positions. The effects of such manipulations are not always perfectly
selective. For example, altering featural information by replacing
the eyes and mouth with those of another face could also change
their spatial relations (configural information; as noted in Rhodes,
Brake, & Atkinson, 1993). Also, Rakover (2002) has indicated
that altering configural information by increasing the inter-eye distance could also indirectly induce a change in featural information.
For instance, the bridge of the nose might appear wider. To exclude such problems, we used scrambling and blurring procedures
in our experiments, which allowed us to investigate the roles of
featural and configural information separately. The experiments reported in this study were modeled after the inter-extra-ortho experiment by Bulthoff and Edelman (1992). In order to generate
the stimuli for these experiments, it was necessary to scramble
faces into their facial components (eyes, mouth, nose, etc.), for
details on scrambling in 2D see Schwaninger et al. (2002). In
the following experiments, we made use of a database of 3D faces
(http://faces.kyb.tuebingen.mpg.de) for which:
Experiments
2.1
2.1.2
We first carried out a one sample t-test (one-tailed) for each of the
13 angles in order to test the group means against chance performance (i.e. d0 = 0). For all 13 test angles, recognition performance differed significantly from chance (p < .05). In a next
step, individually calculated d values were subjected to a twofactor analysis of variance (ANOVA) with condition (extra, inter,
ortho (up), ortho (down)) and amount of rotation (0 , 15 , 30 ,
45 ) as within-subjects factors. Mean values are shown in Figure 2. Recognition d0 was found to be dependent on the condition, as indicated by the main effect of condition F (3, 27) =
21
22.91, p < .001. There was also a main effect of amount of rotation, F (3, 27) = 29.35, p < .001. As illustrated by Figure 2,
the effect of amount of rotation differed across conditions. This is
also indicated by the interaction between amount of rotation and
condition, F (9, 81) = 7.29, p < .001.
2.2.1
2.2
2.2.2
22
ing angle of rotation (see Figure 4). Thus not only whole-face processing, but also featural processing can be explained by a linear
interpolation model. No conclusive answer was found for the question of a symmetry effect in scrambled-face recognition. There was
no significant difference between the inter and the extra conditions.
While Figure 4 shows a clear trend towards such a difference, further investigation is necessary to provide a definite answer. However, Experiment 2 clearly showed that the effect of a horizontal
benefit seems to disappear with scrambled faces.
2.3
23
scrambled, in which all components were cut out and then rearranged
other 12 test angles differed significantly from 0 (p < .05). The results show that human observers are capable of recognizing faces on
the sole basis of configural information for almost all viewpoints.
In a second step, we compared the data with the data from the
first baseline experiment. To do this, we calculated a three-factor
ANOVA with the within-subjects factors of condition (extra, inter, ortho (up), ortho (down)) and amount of rotation (0 , 15 ,
30 , 45 ) and the between-subjects factor of stimulus type (intact face vs. blurred face). We found a main effect of stimulus
type, F (1, 18) = 26.30, p < .001 and main effects of the withinfactors condition F (3, 54) = 44.16, p < .001 and amount of rotation F (1.66, 29.86) = 25.20, p < .001. There was also an interaction effect between stimulus type and condition, F (3, 54) =
6.06, p < .01, but no interaction effect between stimulus type and
amount of rotation, F (1.66, 29.86) = 1.08, p > .34. We also
found an interaction effect between amount of rotation and condition F (4.60, 82.71) = 12.52, p < .001 and a three-way interaction between amount of rotation, condition and stimulus type
F (4.60, 82.71) = 2.78, p < .05. A paired-sample t-test between the two conditions (whole face vs. blurred face) showed a
significantly better overall recognition performance for the wholeface testing condition than for the blurred-face testing condition
(p < .001).
scrambled-blurred, in which the same low-pass filter was applied to the scrambled components
The blur-level had been determined in previous testing by determining the blurring strength at which performance in the scrambledblurred condition was at chance. The idea behind this manipulation
was that in the scrambled condition, all configural (and holistic) information was disrupted so that faces could only be recognized on
the basis of individual components. When these rearranged components were blurred, participants were simply guessing, meaning
that all discriminative information from the components was removed. When applied to the intact faces, this level of blur therefore removes information about fine detail required for component
processing but retains the first- and, especially, second-order relational information (Diamond & Carey, 1986) thought to be the
basis for configural face processing (Collishaw & Hole, 2000; Sergent, 1984). Schwaninger et al. (2002) showed that recognition of
both scrambled (component-based) and blurred (configural-based)
faces was above chance, adding to previous evidence that routes to
recognition exist for both types of information. In addition, both
yielded better performance for familiar than unfamiliar faces.
2.3.1
In a next stage, we calculated paired-sample t-tests between wholeface testing condition, scrambled testing condition, and blurred
testing condition. The results show a significantly better overall recognition performance for the whole-face testing condition
than for the blurred face testing condition (p < .001) and a significantly better overall recognition performance for the blurredface testing condition than for the scrambled-face testing condition (p < .001). We also calculated a two-factor ANOVA including the factors of condition (extra, inter, ortho (up), ortho (down))
and amount of rotation (0 , 15 , 30 , 45 ), in order to focus on
the qualitative characteristics of configural processing. Again, we
took into account only the data from the blurred condition. In this
case, main effects of condition (inter, extra, ortho (right) or ortho
(left)) F (1.57, 14.17) = 25.98, p < .001 and amount of rotation
(0 , 15 , 30 , 45 ) F (1.50, 13.48) = 6.89, p < .05 were also
found. There was also an interaction between the amount of rotation and condition F (3.50, 31.46) = 7.96, p < .001. A pairwisecomparison t-test between the four conditions showed that recognition performance in the inter condition was significantly better than
in the other three conditions (extra p < .05; ortho p < .001). There
was no significant difference between recognition performances in
the two ortho conditions (p > .11). The extra condition showed a
significantly better recognition performance than the ortho (down)
condition (p < .001), but did not differ significantly form the ortho
(up) condition (p > .09).
In summary, it is clear that human observers are capable of recognizing faces across different viewpoints on the sole basis of isolated
configural information. This supports the idea that configural information is important for face recognition and that it can also be processed independently of other information. A comparison between
the first two experiments shows that isolated configural information
leads to a considerably lower recognition performance than wholeface processing but also to considerably higher recognition performance than featural processing. This finding is consistent with that
of Schwaninger et al. (2002), who also reported a better recognition
performance for configural processing than for featural processing.
The current results confirm that the human perceptual system is
capable of recognizing faces on the basis of isolated information
such as configural or featural information, but they also demonstrate that integration from different information sources leads to
the best recognition performance. Again, we found a systematic effect of viewpoint, in that recognition performance increased with a
24
decreasing angle between a learned view and a tested view. As Figure 5 illustrates, there is enhanced recognition performance for the
opposite 45 view similar to the results found in Experiment 1 using whole intact faces. Since this effect disappeared for scrambled
faces in Experiment 2, we can conclude that view generalization for
symmetrical views relies on configural information.
General Discussion
Another interesting finding from our experiments is that the symmetry effect and the range of canonical views along the horizontal
meridian were only found in connection with configural information, but not with featural information. This is consistent with the
assumption that configural information is used as basic information
for initial classification, and that featural information is used for
further differentiation between individuals. In everyday life, faces
are seen mostly from frontal or side views. Up or down views are
less common, and hence less important. For the visual system, it
is important to make a quick and reliable estimate in order to assess whether the object is relevant or whether further information
is needed - this could be done by the configural processing route
as we found good generalization performance along the horizontal
axis. For featural processes, such a benefit for the horizontal axis
is less essential as this type of processing might only be required
when further, more detailed information is needed.
Acknowledgements
We would like to thank Jenny Campos and three anonymous
reviewers for valuable comments on an earlier draft of this
manuscript.
25
References
emotion in faces: A psychophysical, physiological and computational perspective. In Progress in Brain Research, vol. 156,
321343.
S CHWANINGER , A., AND M AST, F. 1999. Why is face recognition so orientation-sensitive? psychophysical evidence for an
integrative model. In Perception (Suppl.), vol. 28, 116.
B ULTHOFF
, H. H., AND E DELMANN , S. 1992. A morphable
model for the synthesis of 3d faces. In Proceedings of the National Academy of Sciences of the United States of America,
vol. 89(1), 6064.
S CHWANINGER , A., C ARBON , C. C., AND L EDER , H. 2003. Expert face processing: Specialization and constraints. In Development of face processing, Gottingen: Hogrefe, G. Schwarzer and
H. Leder, Eds., 8197.
B ULTHOFF
, H. H. submitted. View-based recognition of faces:
Generalization across viewpoint is orientation-sensitive. In Vision Research.
C OLLISHAW, S. M., AND H OLE , G. J. 2000. Featural and configurational processes in the recognition of faces of different familiarity. In Perception, vol. 29(8), 893909.
D IAMOND , R., AND C AREY, S. 1986. Why faces are and are
not special: An effect of expertise. In Journal of Experimental
Psychology: General, vol. 115, 107117.
S ERGENT, J. 1984. An investigation into component and configurational processes underlying face recognition. In British Journal of Psychology, vol. 75, 221242.
H ILL , H., S CHYNS , P. G., AND A KAMATSU , S. 1997. Information and viewpoint dependence in face recognition. In Cognition,
vol. 62(2), 201222.
M OSES , Y., U LLMAN , S., AND E DELMAN , S. 1996. Generalization to novel images in upright and inverted faces. In Perception,
vol. 25(4), 443461.
PALMER , S., ROSCH , E., AND C HASE , P. 1981. Canonical perspective and the perception of objects. In Attention and Performances IX, Lawrence Erlbaum: Hillsdale, NJ, J. L. . A. Baddeley, Ed.
U LLMAN , S., AND BASRI , R. 1991. Recognition by linearcombinations of models. In IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 13(10), 9921006.
B ULTHOFF
, H. H. 2002. View-based recognition of faces in man
and machine: Re-visiting inter-extra-ortho. In Lectures Notes in
Computer Science, vol. 2525, 651660.
26