Adapted active appearance models

被引:0
|
作者
Seguier, Renaud [1 ]
Le Gallou, Sylvain [2 ]
Breton, Gaspard [2 ]
Garcia, Christophe [2 ]
机构
[1] IETR, SUPELEC, F-35511 Cesson Sevigne, France
[2] IRIS, Orange Labs TECH, F-35512 Cesson Sevigne, France
关键词
Active Appearance Model; Human Machine Interface; Face analysis;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Active Appearance Models (AAM) are able to align known faces in an efficient manner when face pose and illumination are controlled. The AAM exploit a set of face examples in order to extract a statistical model. There is no difficulty to align a face with the same type (same morphology, illumination and pose) which constitute the example data set. Unfortunately, the AAM are less outstanding from the moment when the illumination, pose and face type changes. AAM robustness is link to the variability introduced in the learning base. The more the AAM will contain variability, the more it will be able to adapt itself to variable faces with the following drawback: the data represented in the reduced parameters space will then form different classes letting appear holes, regions without any data (see Fig. 1). It is therefore very difficult to make the AAM converge in this scattered space. We propose in this paper a robust Active Appearance Models allowing a real-time implementation. To increase the AAM robustness to illumination changes, we propose Oriented Map AAM (OM-AAM). Adapted AAM will be presented after to increase the AAM robustness to any other types of variability (in identity, pose, expression etc.). OM-AAM We propose a specific transformation of the active model texture in an orientation map, which change the normalization process of the AAM. First of all, we apply systematically an adaptive histogram equalization from CLAHE [Zuiderweld 1994] on the images. We then evaluate the horizontal and vertical gradient and simply use the gradient angle on each pixel instead of its gray level. This angle is evaluated on N-a values. In practice we quantify it on eight bits so N-a = 255. The new texture is then made up from an image representing the orientation of each pixel, that we call an oriented map. To overcome a discontinuity problem between 0 and 2Pi associated to similar edge orientations, we realize a mapping (Eq. 11) from [0..2 Pi] to [0..Pi/2]. In order to reduce the noise in uniform regions as illustrated in the background of Figure 3c, we emphasize the signal correlated with the region with high gradient information as it is proposed by [Cooste 2001] in using a non-linear function f (see Eq. 12). During the modelization, the oriented texture from image 14 (Eq. 13) will replace the texture usually used by the AAM. We compare the OM-AAM performance to those of the DM-AAM and classical AAM on two public databases. The first one is dedicated to illumination problems (CMU-PIE: 1386 images of 66 faces under 21 different illuminations [SIM 2002]) and the other one is composed of different faces with several expressions taken in different backgrounds (BIOID: 1521 images [Res01]) under variable light exposition (see Fig. 5). These comparisons will be made in a generalization context: the faces used to construct the model (18 persons from the M2VTS database [Pigeon 1996]) do not belong to the testing databases. We normalize the error found on four relevant points (gravity centers of the eyes, nose and mouth) by the distance between the eyes (see Eq. 15). Figure 6 represents the percentage of the images which have been aligned with the error e. On the CMU-PIE database, the OM-AAM are able to align 94% of the faces with a precision less or equal to 15% where the DM-AAM and classical ones are less efficient: their performances are of 88% and 79% respectively. But when the faces are acquired in real situations, our proposition overcome other methods: on BIOID database, the OM-AAM can align 52% of the faces with a precision less or equal to 15% which represents a gain of 27 and 42% in the performance relatively to classical AAM and DM ones respectively. Adapted AAM As said above, the AAM robustness is related to the face variability in the learning base. Instead of using a very generic model containing a lot of variability, we propose to use an initial model M-o which contains only a variability in identity and then use a specific model M-adapt containing variability in pose and expression. Let a general database contains three types of variability: in expression, identity and pose (see Fig. 7). It is made of several different faces, holding four distinct expressions. Each of the face presents each expression in five poses. The initial model Mo is realized from a database BDDo containing different frontal faces in neutral expression (see Fig. 8). This initial model will be used to perform a rough alignment on the unknown face. Let C-o be the appearance vector after the alignment of the model Mo on the unknown analyzed face. In the reduced space of the model parameters, we seek for the k nearest parameters vectors of Co belonging to the learning initial database BDDo. These k nearest neighbors correspond to the k nearest face of the analyzed one. For example in figure 9, the vector CP will identify the face number p as the most resemble to the analyzed one. From this set of k nearest identities, we generate an adapted database BDDadapt containing the corresponding faces in different expressions and poses, BDDadapt is a subset of the general database (Fig. 7) from which we generate the adapted model M-adapt. When k = 1, 2 or 3 it is possible to evaluate beforehand the adapted model. When we need to align an unknown face in a static image, we then simply align the face with the initial model Mo and apply the pre-computed model which corresponds to the k nearest faces. We test the adapted AAM on the static images of the general database B D Do (Fig. 7). A test sequence is then constituted of one unknown person showing four expressions under five different poses, the learning base associated to this testing base being constituted of all the other persons. A cross-validation of type Leave-one-out is used: every face is tested separately, using all the other one for the learning base. All the faces of the database have been tested, representing at the end a set of 580 images. We compare the performance of our system when k = 2 ("Adapted AAM") to the two other different AAM. The first one ("AAM 28") get identity as the only variability and is constructed from the 28 faces (the twenty-ninth being tested) in frontal view and neutral expression. The second one ("AAM 560") is full variability since it is based on 560 images representing 28 faces showing four expressions under five different poses. Figure 11 shows the superiority of the "Adapted AAM" on the two other models. If we look at the performances at the reference error of 15% our proposition is ten times more rapid and 16% more effective than a model constructed from a more rich database ("AAM 560"). If now we compare the "Adapted AAM" to the "AAM 28" which has the same computational complexity, it is more effective in 55% of the cases (94% versus 49%). As a conclusion our model is more rapid and effective compared to the other models because it focuses on a relevant database relatively to the testing face.
引用
收藏
页码:367 / 380
页数:14
相关论文
共 50 条
  • [21] Gabor texture in active appearance models
    Gao, Xinbo
    Su, Ya
    Li, Xuelong
    Tao, Dacheng
    NEUROCOMPUTING, 2009, 72 (13-15) : 3174 - 3181
  • [22] Increasing the density of active appearance models
    Ramnath, Krishnan
    Baker, Simon
    Matthews, Lain
    Ramanan, Deva
    2008 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, VOLS 1-12, 2008, : 1357 - +
  • [23] Multi Subspaces Active Appearance Models
    Yang, Junyeong
    Byun, Hyeran
    2008 8TH IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE & GESTURE RECOGNITION (FG 2008), VOLS 1 AND 2, 2008, : 122 - 127
  • [24] A comparative study of facial appearance modeling methods for active appearance models
    Lee, Sung Joo
    Park, Kang Ryoung
    Kim, Jaihie
    PATTERN RECOGNITION LETTERS, 2009, 30 (14) : 1335 - 1346
  • [25] Accurate Regression Procedures for Active Appearance Models
    Sauer, Patrick
    Cootes, Tim
    Taylor, Chris
    PROCEEDINGS OF THE BRITISH MACHINE VISION CONFERENCE 2011, 2011,
  • [26] Robust Active Appearance Models Search Algorithm
    Lin, Yong-Fong
    Tang, Chih-Wei
    ADVANCES IN MULTIMEDIA INFORMATION PROCESSING - PCM 2008, 9TH PACIFIC RIM CONFERENCE ON MULTIMEDIA, 2008, 5353 : 879 - 882
  • [27] Active Appearance Models with Rotation Invariant Kernels
    Hamsici, Onur C.
    Martinez, Aleix M.
    2009 IEEE 12TH INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2009, : 1003 - 1009
  • [28] Probabilistic localization by appearance models and active vision
    Kröse, BJA
    Bunschoten, R
    ICRA '99: IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, VOLS 1-4, PROCEEDINGS, 1999, : 2255 - 2260
  • [29] View-based active appearance models
    Cootes, TF
    Wheeler, GV
    Walker, KN
    Taylor, CJ
    IMAGE AND VISION COMPUTING, 2002, 20 (9-10) : 657 - 664
  • [30] Visual tracking using active appearance models
    Birkbeck, N
    Jagersand, M
    1ST CANADIAN CONFERENCE ON COMPUTER AND ROBOT VISION, PROCEEDINGS, 2004, : 2 - 9