• No se han encontrado resultados

Facial action unit detection with convolutional neural networks

N/A
N/A
Protected

Academic year: 2020

Share "Facial action unit detection with convolutional neural networks"

Copied!
19
0
0

Texto completo

(1)

Facial Action Unit Detection with

Convolutional Neural Networks

Andr ´es Romero,

Graduate Student,

Pablo Arbel ´aez,

Assistant Professor

Abstract—We propose a novel deep convolutional neural network architecture to study the problem of action unit detection. We leverage recent gains in large-scale object recognition by formulating the task of predicting the presence of a specific action unit in a still image as simple image-level binary classification. We first train a convolutional encoder on the problem of multi-view emotion recognition as a high-level representation of facial expressions. We show that our architecture generalizes across views, ethnicity, gender and age by merging and training jointly on three standard emotion recognition datasets: CK+, Bosphorus and RafD. Our system is the first fully multi-view emotion recognizer proposed in the literature. We then extend this shared learned representation with fully-connected layers trained to detect individual action units. Our approach is conceptually simpler and yet significantly more accurate than the best methods based on the dominant paradigm for the study of this problem, which relies on facial landmark detection as an intermediate task. We conduct experiments on the BP4D dataset, the largest and most challenging benchmark currently available for action unit detection, and report an absolute improvement of16%over the previous state-of-the-art.

Index Terms—Deep learning, Fine-grained localization, Action Units.

F

1

I

NTRODUCTION

H

UMANfacial expression interpretation has been a clas-sic field of study in psychology, and it has benefited from seminal contributions by renowned researchers such as P. Ekman, who characterized and studied the manifes-tation of prototypical emotions through changes in facial features [1]. From the computer vision perspective, solving the problem of automated facial expression interpretation is a cornerstone towards high-level human computer interac-tion, and its study has become an active topic of research in the last decades. Traditionally, the problem has been approached by focusing on its most basic version and classi-fying static images or short sequences of faces into a handful of prototypical emotions (e.g., happiness, sadness, fear, etc.). Recent methods have achieved significant progress in the study of this task, leading to the near saturation of standard benchmarks such as the CK+ dataset [2].

In order to study facial expressions in a systematic way, Ekman and his collaborators designed the Facial Action Coding System (FACS) [3]. FACS relies on identifying visible local appearance variations in the human face, called Action Units (AUs), produced by to contraction or relaxation in any of its 30 muscles (e.g., a raised eyebrow). AUs constitute therefore a natural physiological basis for face analysis, in which any facial expression can be represented by their combinations.

The existence of a physiological basis for a computer vision domain is a rare luxury, as it allows focusing on the essential atoms of the problem and, by vertue of their multi-ple possible combinations, opens the door to a wide range of applications beyond emotion classification. Consequently, in the last years the main focus of the community has shifted

• Andr´es Felipe Romero Vergara is candidate to the title of Msc in Biomedical Engineering from the Department of Biomedical Engineering, Universidad de Los Andes, Bogot´a, Colombia.

E-mail: rv.andres10@uniandes.edu.co

towards the detection of Action Units, and recent datasets such as BP4D [4] include this type of annotations produced by trained psychologists. The dominant approach for AU detection [5], [6], [7], [8] is divided into two stages: first, facial landmarks are localized in the image and, then, the appearance of patches centered at each keypoint, as well as their relationships, are analyzed in order to determine the presence or absence of each action unit in the image.

In this paper, we depart from the current mainstream approach for action unit detection and bring the power of Convolutional Neural Networks (CNNs), the family of techniques that has revolutionized visual recognition in the last years, to the study of this problem. In the context of automated face analysis, CNNs have been used successfully for emotion classification [9], facial landmark detection [10] and person re-identification [11]. However, to the best of our knowledge, CNNs have never been applied before to the challenging problem of action unit detection, and the current state-of-the-art-systems [5], [6], [7], [8] rely on hand-engineered features.

Figure1presents an overview of our architecture, which we call Convolutional Action Unit Detection (CAUD). First, we train a convolutional encoder for facial expression analy-sis by specializing a network that was originally designed to recognize object categories. We enforce generalization across views, ethnicity, gender and age by training jointly on three leading emotion recognition datasets with very different characteristics: CK+, Bosphorus and RafD. Then, starting directly from this high-level representation, we train fully-connected detectors to determine the presence or absence of each action unit in the image.

CAUD is simpler and yet significantly more powerful than existing action unit detection systems. First, it bypasses keypoint detection, which is a nontrivial problem requiring expensive annotations, and it is trained end-to-end from image pixels to the final goal. Second, it takes advantage

(2)

Fig. 1.Overview of Convolutional Action Unit Detection. A shared convolutional encoder, trained for the task of multi-view emotion classi-fication is first applied to the input face image. Then, specialized fully-connected layers predict the presence in the image of each of the action units considered.

of the information in the whole image by leveraging a powerful learned representation, which is shared among AU detectors. Third, it is efficient at test time, as it requires a single forward pass of the network to detect all AUs. Last, but not least, as the experiments will show, it is modular and leverages better architectures for improved performance, having thus the potential to incorporate future developments in CNNs for general object classification.

We perform an extensive empirical validation on four public datasets, showing that CAUD provides large gains in performance over the previous state-of-the-art. On the BP4D dataset, the largest and most varied available benchmark for action unit detection, we obtain an absolute improvement of16%over the state-of-the-art method of Zhenget al.[7]. Given the absence of publicly available action unit detection systems to use as baseline, we plan to contribute to future research on facial expression analysis by releasing all our source code, benchmarks and results.

2

B

ACKGROUND

Convolutional Neural Networks (CNNs) are a branch of machine learning (supervised classification) based on a set of algorithms that attempt to model high-level abstractions in data by using a deep graph with multiple processing layers, composed of multiple non-linear transformations. This solution allows computers to learn from experience and understand the world in terms of a hierarchy of concepts. By gathering knowledge from experience, this approach avoids the need for human operators to formally specify all of the knowledge that the computer needs. The hierarchy of concepts allows the computer to learn complicated concepts by building them out of simpler ones [12]. Thus, each layer is composed with families of convolutional filters plus a no-linearity step. Therefore, a convolutional neural network aims at learning millions of parameters and consequently huge amounts of data are required to learn such parameters. Fig.2. shows an example of a neural net during train and inference.

Fig. 2.Convolutional Neural Networks [13].Up:Training a CNN. Each time a group of images are passed forward through the network, all parameters are updated in a backward step according to the output loss error function.Bottom:During inference, only images are passed through the network in order to get a desired prediction.

3

R

ELATED

W

ORK

We focus our review of related work on the two main problems of facial expression analysis: emotion classification and action unit detection.

3.1 Emotion Classification

Human emotion recognition has been a challenge for scien-tists for many years. Since P. Ekman [1] categorized emo-tions into six classes (happiness, sadness, anger, surprise, fear and disgust), this problem has been studied using dif-ferent image representations, such as LBP [14], [15], Weber local descriptors (WLD) [16] and SIFT [17], and commonly used classifiers such as Support Vector Machines (SVM) [18] or Bayesian networks [19].

Recently, convolutional networks have been applied to emotion classification from still images [9], [20], [21], and recurrent networks have been used to model time depen-dencies for emotion recognition in videos [22], [23]. Not surprisingly, CNN-based approaches achieve state-of-the-art performance on public benchmarks, and surpass [24] human performance on CK+ [2], the traditional dataset for emotion classification. Nevertheless, the final goal of these techniques is the classification problem by itself, while we propose to use emotion classification to learn a shared representation for subsequent action unit detection.

Traditionally, this problem has been studied using only frontal face images [2], [25], [26], [27], [28]. Although RafD [29], a multi-view dataset, has been annotated with emotions, existing approaches that use it as a testbed [30], [31], [32] consider only frontal faces and ignore the other views. As our intention is to build a general and robust system for emotion recognition, rather than to improve state-of-the-art performance on a specific database, one of our contributions is to explore the generalization capability of CNN-based approaches to multiple views and datasets.

3.2 Action Unit Detection

Figure3presents examples of the 12 action units considered in the experiments. As can be observed, action unit detection is a challenging fine-grained recognition task.

Existing methods to action unit detection [5], [6], [7], [8], all follow the paradigm of first detecting facial landmarks using Active Appearance Models [33], and then treating

(3)

Fig. 3.Action Unit Examples. Two examples are presented for each one of the action units considered in this paper, cropping the region of the face where the action unit is expressed. [. 3]

these keypoints as anchors to assess the presence of specific action units in the image. Thus, Zhao et al. [8] use facial keypoint detectors as a pre-processing tool for this task, then extract patches around keypoints and measure co-occurrences between action units to train binary Support Vector Machine classifiers. Similarly, Zenget al.[7] start from facial keypoints detector, extract SIFT descriptors around keypoints, and then use a quasi-semi-supervised approach to train a pair of classifiers for positive and negative sam-ples.

We depart radically from the current paradigm on action unit detection and propose to use CNNs directly on the original image, in order to learn automatically a high-level representation for facial expressions, which is then special-ized to detect the presence of individual action units in the image.

4

C

ONVOLUTIONAL

A

CTION

U

NIT

D

ETECTION

This section describes our deep architecture for Action Unit detection. We first train a classification network for the task of general emotion recognition from multiple views of the face, and then we use its convolutional layers as input for training fully connected layers that are specialized in detecting specific Action Units.

4.1 Emotions

Recent works have shown that starting from a convolutional encoder, trained on a large-scale classification problem such as Imagenet, can be useful for different applications, such as contour detection [34], [35] or semantic segmentation [36]. We are inspired by these approaches to learn a convolutional encoder based on emotion recognition. Consequently, we aim at developing the most general and robust system capable of extracting the best features based not only on frontal images of the face, but on multiple views of the face.

Fig. 4. Individual Action Unit Detection. By formulating action unit detection as a simple binary classification task on the whole image, we can transfer recent progress in large-scale object recognition to the domain of facial expression analysis, and specialize networks pre-trained in Imagenet for the task of predicting the presence or absence of a specific AU.

Additionally, Khorrami et al. [37], verified that lower layers of a CNN trained on emotion classification indeed learn portions of the face associated to certain emotions. In other words, lower layers of a CNN have the ability to learn the main basis of AUs. Therefore, we propose to build our emotion classification model, using 3 different datasets (CK+ [2], Bosphorus [28], and RafD [29]). By considering them jointly, our system learns to generalize across views (front and five camera angles), color (RGB and Gray), differ-ent resolutions and scales.

We start from a deep architecture trained on Imagenet and modify the output layer to predict 8 emotions, includ-ing the ’neutral’ category, which is the most common and challenging one. Then we fine-tune the network on our multi-view dataset for emotion classification, optimizing with Stochastic Gradient Descent (SGD) [38].

To the best of our knowledge, there is no other existing emotion classification model that takes into account multi-view perspective. Our model can be used as baseline for future reference.

4.2 Action Units

Our approach is motivated by the observation that detect-ing a specific action unit, i.e. determining its presence or absence in the image, can be formulated as a standard binary classification task for the whole image. Therefore, we adapt recent architectures that have been designed for image classification in Imagenet to our problem domain.

To do so, we start from the multi-view emotion recog-nition model with convolutional encoder layers frozen, we change the output layer only to binary classification (pres-ence or abs(pres-ence) and then fine-tune in order to specialize the fully connected layers. Each Action Unit is trained in isolation until convergence, then it is done iteratively for the 12 Action Units considered. Figure4shows a representation on the network for each Action Unit.

Once each Action Unit is trained, we can simply join them together in a big network architecture CAUD. This can be easily performed thanks to the shared weights in the convolutional encoder layers, common in all of them, as seen in Figure1. This last reason is fundamental, if it is not performed, training Action Units on different networks would not be feasible in practical terms.

Our approach is simpler and more accurate than existing systems. It does not rely on external methods for keypoint detection. It is completely trainable end-to-end. Its modular

(4)

nature is undoubtedly one of its biggest advantages, because it allows training news AUs in isolation and then adding them to the architecture, contributing thus to the generality and robustness of the entire system. It is also efficient at test time, as with a simple forward pass it predicts multiple Action Units from the a raw image.

5

R

ESULTS

We first describe the experimental framework we adopt and report experiments on multi-view emotion classification. Then we present our main results on AU detection, for which we first analyze the different modules of our system and then we compare our result against the state-of-the-art.

5.1 Experimental framework

5.1.1 Multi-view emotion dataset

Traditionally, emotion recognition has been approached by restricting the analysis to images of frontal faces. Since our goal is to use this task in order to learn a generic high-level representation for facial expressions, we require the largest possible diversity in our data. Therefore, for our experiments on emotion classification, we merge three publicly available datasets with very different characteris-tics: CK+ [2], Bosphorus [28] and RafD [29]. The emotions we want to predict are happy, sad, anger, contempt, fear, disgust, and surprise, including neutral. Although other datasets for emotion classification have been proposed [25], [26], [27], we did not include them because they do not take into account the ’contempt’ emotion.

5.1.1.1 CK+: It is a traditional testbed for facial expression analysis. It was constructed by collecting 593 sequences of posed facial actions from 123 subjects. The first and the last frame were selected as negative (neutral) and positive (emotion).

5.1.1.2 Bosphorus: This database is unique in two aspects: (i) the facial expressions were composed by judi-ciously selecting a subset of Action Units (using these, emo-tion labels were annotated), and many actors/actresses were incorporated to obtain more realistic expression data; (ii) a rich set of head pose variations are available, which adds complexity to the data. This dataset provides 453 images properly annotated with emotions from 105 participants.

5.1.1.3 RafD: The Radboud Faces Database, offers a stimulus set containing both Caucasian adult and children images. However, its most important contribution is the emotion recognition not only from frontal faces, but from different camera angles, even with full profile images. It consists of 4098 images from 73 subjects.

In summary, our proposed multi-view dataset contains 5737 images from 301 subjects with large variations in pose, age and ethnicity of subjects. It contains both gray-level and color images at multiple resolutions.

Additionally, for Action Unit detection, we use BP4D [4], [39], the largest and most challenging dataset available for this task.

5.1.2 BP4D

BP4D-Spontaneous database consists of 2D/3D videos of spontaneous facial expressions in young people, all these videos were recorded in non-acted scenarios, since all emo-tions were elicited by experts in the subject. This dataset provides around 60,000 frames from 101 subjects properly annotated by trained psychologists with action units.

5.1.3 Evaluation metrics

We report results using the standard performance metrics from the literature. For emotions classification we report Average Classification Accuracy (ACA) [40], which is the average of the diagonal on the normalized confusion matrix. For Action Units, we evaluate performance using image based F1-score (f1-frame) [41], also known as F-measure, which considers the optimal regime on the precision and recall curve of detected AUs. F1-score (2 · PP·R+R) is the traditional metric in binary classification tasks. Additionally, for AUs, we report the Average Precision (AP) metric, which combines information across all operating regimes.

5.1.4 Dataset Splits

Common practice in the field is to perform k-fold cross validation as evaluation methodology. Although such a framework is justified in the presence of scarce data, the datasets considered in this paper are large enough so that training and validating on2/3 of the data provides mean-ingful results. Furthermore, a unique explicit split of the data allows more transparency and the ability to use the optimal parameters learned for an initial task to be used as input for a second stage of processing. Therefore, in all the experiments apart from the comparison with the state-of-the-art on action unit detection, we use the following split of the data. Each dataset was divided in a subject-exclusive manner, in training, validation and test, with the same number of subjects each. This is mandatory due to the independence added by treating all subjects separately. If images of the same subject appear in different splits, systems might memorize characteristics specific to the people in the dataset rather than learning action units.

In order to compare our method directly with state-of-the-art AU detectors, and set a benchmark for future reference, in the BP4D dataset we also follow the exper-imental framework of [7], [8] and perform 10-fold cross validation using their splits. In all the experiments we use the evaluation code of [7] to ensure comparable numbers.

5.1.5 Implementation Details

The BP4D dataset provides only frontal face images, but they depict torsos and the heads are not full centered. Therefore, as a pre-processing, we apply the Viola-Jones [42] face detector on each image. We restrict ourselves to such a simple external face detector because of the high quality of the data and the unique frontal view of the entire set. Those rare faces that are not correctly detected, are considered as errors of our algorithm in the evaluation.

All the experiments were conducted using the publicly available deep learning frameworkcaffe[43].

(5)

5.1.6 Reproducibility

We would have liked to compare our results against all previous literature on AU detection, but faced the following reproducibility challenges:

• State-of-the-art emotion recognition works [14], [24] evaluate performance by ignoring the ’neutral’ label, which is the most common and difficult one. • There is no agreement on a standard performance

metric in the literature [5], [6], [7], [8] for AU detec-tion.

• We could not find any publicly available baseline for AU detection, nor pre-computed results for compar-ison.

Consequently, we could only establish a direct empirical comparison with the state-of-the-art AU detection method of [7], for which the authors provided the split of the BP4D dataset used in their paper.

In order to promote reproducible research on facial ex-pression analysis, we plan to publicly release all our source code, benchmarks and results.

5.2 Multiview Emotion Classification

In this section, we report experiments on multi-view emotion classification.

We perform data augmentation as it is typical with large capacity CNNs in order to cope with the lack of data. Thus, we perform rotations of several degrees and jittering on the entire set.

We aim at comparing two different pre-trained architectures on Imagenet, the seminal Alexnet [44], and the deeper, more accurate and more performant VGG-16 [45].

We train our model during 10 hours (Alexnet took less time because it has fewer parameters than VGG) using a GPU K40 with 10−4 learning rate,5x10−3 weight decay and 100 epochs.

In this general version of the problem, we do not have an external reference, so we use a standard Bag-of-Words (BoW) [46] approach as our baseline in order to establish the difficulty of the problem and to assess the improvement of deep learning techniques with respect to hand-engineered features.

Table 1 shows the per-category classification accuracy, evaluated on the test set of the multi-view dataset. Additionally, we present on Table 2, the confusion matrix for the VGG model.

In the remainder of this paper, we choose the best model, VGG with an ACA of86%in the all the experiments.

5.3 Action Unit Detection

5.3.1 Control Experiments

We first report control experiments assessing the relative importance of different modules of our system. We train

TABLE 1

Results on multi-view emotion classification.Comparison between two different Imagenet architectures and a bag-of-words baseline. Results are reported with average classification accuracy on the test

set of the multi-view dataset. Best results are shown in bold.

Emotions BoW Ours

Alexnet VGG

Neutral 25,9 70,5 79,9

Happy 41,1 81,8 96,9

Sad 33,0 78,9 84,9

Anger 44,2 76,5 87,2

Fear 28,7 61,2 81,5

Disgust 38,4 82,7 92,5

Surprise 67,9 85,1 90,7

Contempt 15,7 67,1 74,1

Av. 38,1 75,5 86,0

TABLE 2

Confusion matrix for multi-view emotion classification. This confusion matrix are the results of VGG on multi-view test set.

Neutral Happy Sad Anger Fear Disgust Surprise Contempt Neutral 79,92 0,16 8,60 2,46 1,98 1,49 0,20 11,84

Happy 0,00 96,88 0,00 0,00 0,20 1,99 0,00 1,09

Sad 4,95 0,16 84,86 4,38 1,98 0,33 0,00 1,27

Anger 2,06 0,00 6,17 87,19 0,79 1,32 0,00 2,37

Fear 0,96 1,31 5,61 0,17 81,55 0,66 7,71 0,73

Disgust 2,34 1,80 0,93 1,05 0,99 92,55 0,20 0,00

Surprise 0,00 0,00 0,00 0,00 9,29 0,00 90,71 0,00

Contempt 11,42 0,98 2,43 7,02 0,00 0,00 0,00 74,13

12 different binary classifiers for 12 different AUs, freezing convolutional encoder layers, and then we fuse them into one big network. Recall that this can be achieved thanks to the shared weights on convolutional encoder layers.

First, we want to analyze the difference in performance when using two different pre-trained models, one model trained on the Imagenet challenge [45] and our model trained on emotions, which we call EmotionNet. These models are applied to the VGG-16 network since it gets the best performance among EmotionNets. In Table 3 we summarize the results on each pre-trained model, using two different metrics.

As a result of joining several fully connected layers, the amount of parameters of this networks is very large, so we need to assess the importance of having so many parameters. We proceed to analyze 3 different subsets with fully connected parameters on: 4096, 2048, and 1024.The F1-frame and Average Precision (AP) are reported on Table 4. These scores were obtained by training on 2/3 of the dataset and evaluating on the remaining 1/3, as specified in Section5.1.4.

Figure 6 presents qualitative results on Action Unit detection.

5.3.2 Comparison with iCMP [7]

We now compare our results against the state-of-the-art approach of [7] using 10-fold cross validation with the same dataset splits and evaluation software.

(6)

Fig. 5.Qualitative Results for Multi-view Emotion Recognition. Top two predictions for each image are displayed, sorted by classification score. Correct categories are shown in green and incorrect categories are shown in red. Observe that, mistakes occur in difficult cases, for those the second prediction of our system is generally correct. See AppendixAfor further results.

TABLE 3

Control experiments for convolutional encoder. Results are reported using f1-frame and AP metrics on the third part of BP4D as

suggested in Section5.1.4.

AU F1-frame AP

Imagenet EmotionNet Imagenet EmotionNet

1 31,5 32,1 75,8 74,9

2 30,3 46,5 79,7 84,6

4 44,0 64,0 75,1 85,2

6 80,5 81,2 78,3 80,4

7 80,3 70,2 72,3 65,0

10 80,9 83,9 81,1 80,2

12 87,9 88,4 86,2 86,3

14 62,8 60,1 62,0 62,4

15 53,3 62,6 71,2 78,8

17 65,6 63,5 68,3 66,1

23 35,3 44,1 65,6 74,5

24 49,4 54,6 67,6 75,8

Av. 58,5 62,61 73,6 76,2

Table5summarizes the average of 10-fold cross valida-tion for each acvalida-tion unit.

6

D

ISCUSSION

6.1 Multiview Emotion Classification

The Bag-of-Words model obtains only38%of performance, suggesting the difficulty of this fine-grained recognition

TABLE 4

Control experiments for fully-connected (FC) specialized AU detectors. FC layer in VGG network is modified, trained and tested for

each parameter.

AU F1-frame AP

FC-4096 FC-2048 FC-1024 FC-4096 FC-2048 FC-1024 1 32,1 28,2 25,4 74,9 73,5 66,8 2 46,5 39,6 41,3 84,6 69,6 72,4 4 64,0 65,6 62,7 85,2 85,1 84,0 6 81,2 76,6 73,1 80,4 77,0 75,7 7 70,2 70,4 68,2 65,0 64,8 62,5 10 83,9 81,8 70,8 80,2 78,6 68,9 12 88,4 88,3 84,3 86,3 86,1 82,9 14 60,1 52,6 56,3 62,4 60,7 61,2 15 62,6 53,7 48,7 78,8 70,7 65,8 17 63,6 63,5 62,1 66,1 62,2 63,5 23 44,1 42,2 35,6 74,5 69,8 72,6 24 54,6 49,4 56,0 75,8 69,2 75,0

Av. 62,6 59,3 57,0 76,2 72,7 70,1

Inner Brow Raiser (AU1) Outer Brow Raiser (AU2) Brow Lowerer (AU4) Cheek Raiser (AU6) Lid Tightener (AU7) Upper Lip Raiser (AU10) Lip Corner Puller (AU12) Dimpler (AU14) Lip Corner Depressor (AU15) Chin Raiser (AU17) Lip Tightener (AU23) True Positives False Positives False Negatives Lip Pressor (AU24)

Fig. 6.Qualitative results for Action Unit detection. CAUD fails on difficult images but it succeed on most of Action Units. Examples on 12 Action Units are displayed. Additional results are available on Appendix

B.

task. However, deep learning techniques double the BoW performance, showing the advantage of end-to-end trainable techniques over standard methods in which feature extraction and classification are independent steps. Additionally, we observe that deeper and better Imagenet architectures translate into improved emotion recognition results, which indicates that our method can potentially benefit from current advances in large-scale classification such as ResNet [47].

According to Fig. 2. It is worth noting that confusions are symmetric with respect to the diagonal, e.g. the most common mis-classification for surprise category is fear, and the most common mis-classification for fear is surprise, so the model fails on emotions that are closely related. Most common confusion (11.5%) us between ”Neutral” and ”Contempt”, the categories ignored in most of the literature.

(7)

TABLE 5

10-fold cross validation for each Action Unit.Comparisons with state-of-the-art method and additional methods reported by [7].

AU SVM Ada Lap DAM MDA GFK iCPM [7] Ours 1 46,0 41,5 43,8 38,2 39,6 42,4 46,6 67,3 2 38,5 12,4 17,6 27,3 37,0 35,8 38,7 59,1 4 48,5 39,4 27,2 29,1 45,7 47,3 46,5 69,5 6 67,0 71,7 71,5 67,5 69,2 71,2 68,4 78,4 7 72,2 74,7 71,6 72,6 70,2 72,5 73,8 84,4 10 72,7 75,7 72,8 74,4 71,0 74,2 74,1 85,0 12 83,6 84,3 84,3 76,4 81,8 83,9 84,6 84,9 14 59,9 61,0 62,6 59,9 57,8 57,2 62,2 72,7 15 41,1 30,6 35,2 15,9 41,4 40,6 44,3 62,6 17 55,6 56,6 59,1 52,9 50,1 55,4 57,5 75,8 23 40,8 33,0 33,6 3,9 36,2 39,9 41,7 59,0 24 42,1 34,2 40,5 4,9 41,1 41,7 39,7 67,0 Av. 55,7 51,3 54,7 42,6 53,4 55,2 56,1 72,6

Figure5presents qualitative results on multi-view emotion classification. Results on relatively easy images are good as expected, and the interesting information comes from confusions in difficult cases. False positives are either very hard images to classify, or ambiguous, but in 100% of cases, the second best choice, according to the histogram of confidence, is the correct label. This behavior dictates that the model is learning to focus on specific facial regions, which motivates its use as starting-point to classify action units, and not in the opposite way.

We have developed a robust and general model which we can use as starting-point for Action Units detection. We aim now at build a more robust system by using images in the wild since the databases we used, they were built in constraint laboratory conditions e.g., the same shirt for RafD, faces centered in Bosphorus and posed emotions in all of them. Emotions in the wild definitively improve our starting-point for AU detection.

6.2 Action Unit Detection

According to Table 3, using a multi-view emotion recognition system as pre-trained model for convolutional encoder layers works better for our final task. Both results are close to each other but, as expected, EmotionNet is a better convolutional encoder for action unit detection. It is worth noting that training the model from scratch (random diverge) does not converge, signaling thus the importance of starting from a pre-trained model.

Furthermore, We observe on Table 4 a steady decline in performance on both metrics as the number of fully-connected layers is decreased. Depending on the application and hardware’s memory, one can reduce the number of fully connected parameters without affecting performance significantly.

Figure 6 presents qualitative results on Action Unit detection. These results are very promising, because they show that our model is detecting challenging cases with high accuracy. However, as for multi-view emotion classification, the important analysis comes from those cases

where the model fails, since they provide insights on the facial features the model is learning to analyze expressions in still images.

Our final discussion is related to our biggest contribution in this area, in Table5We observe that CAUD outperforms consistently iCPM for all action units by a significant margin. The average improvement is 16% in f1-score, pushing the state-of-the-art to a new level. To the best of our knowledge, our method is the first to successfully classify multiple facial micro expressions in a complete end-to-end trainable fashion. Moreover, our architecture is modular, i.e. new action units can be trained in isolation and added to the current network.

7

C

ONCLUSIONS

We have presented Convolutional Action Unit Detection for recognizing 12 different facial micro-expressions in still images. Our method takes advantage of the power of CNNs for large-scale classification problems, and is capable of detecting multiple Action Units simultaneously. CAUD leverages a convolutional encoder learned from multi-view emotion classification in order to detect action units. At its core lies a flexible and modular architecture that can incorporate new action units.

CAUD compares favorably against the state-of-the-art, and provides a large improvement in performance in a challenging benchmark.

The number of applications for this solution are unimaginable, every person could give it a different approach according to his/her formation basis i.e., for a Psychologist it could help him to study the relations between Attention-deficit/hyperactivity disorder (ADHD) with facial expressions; for a Physician it serve as aid for depression diagnosis; It could bring to a new level the intelligence personal assistants, among many other disciplines.

A

CKNOWLEDGMENTS

I would like to thank Google for the Google Research Grant from Google Research Awards for Latin America, for the sponsorship during the last year of the project development, and NVIDIA for providing K40 GPU from the NVIDIA’s Academic Hardware Grant Program.

Personally I would like to thank to my parents for never stop believing in my capabilities and remind me every day that the world is a small place to live; my siblings who always have been there every time I need them, and encourage me to overcome regular thinking; my girlfriend who never give up with me, even when the results were not good enough and my mood and facial expressions were not the best to offer her.

Special thanks to Pablo Arbel´aez, without him, none of this would have been possible, with his guidance I am quite sure, something big can be cooked.

(8)

R

EFERENCES

[1] P. Ekman and W. V. Friesen, “Facial action coding system,” 1977. [2] P. Lucey, J. F. Cohn, T. Kanade, J. Saragih, Z. Ambadar, and

I. Matthews, “The extended cohn-kanade dataset (ck+): A com-plete dataset for action unit and emotion-specified expression,” in CVPR 2010. IEEE, 2010, pp. 94–101.

[3] P. Ekman and E. L. Rosenberg,What the face reveals: Basic and applied studies of spontaneous expression using the Facial Action Coding System (FACS). Oxford University Press, USA, 1997.

[4] X. Zhang, L. Yin, J. F. Cohn, S. Canavan, M. Reale, A. Horowitz, P. Liu, and J. M. Girard, “Bp4d-spontaneous: a high-resolution spontaneous 3d dynamic facial expression database,” Image and Vision Computing, vol. 32, no. 10, pp. 692–706, 2014.

[5] S. Kaltwang, S. Todorovic, and M. Pantic, “Latent trees for esti-mating intensity of facial action units,” inProceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2015, pp. 296–304.

[6] T. Almaev, B. Martinez, and M. Valstar, “Learning to transfer: transferring latent task structures and its application to person-specific facial action unit detection,” in Proceedings of the IEEE International Conference on Computer Vision, 2015, pp. 3774–3782. [7] J. Zeng, W.-S. Chu, F. De la Torre, J. F. Cohn, and Z. Xiong,

“Confidence preserving machine for facial action unit detection,” inICCV 2015, 2015, pp. 3622–3630.

[8] K. Zhao, W.-S. Chu, F. De la Torre, J. F. Cohn, and H. Zhang, “Joint patch and multi-label learning for facial action unit detection,” in CVPR 2015, 2015, pp. 2207–2216.

[9] B.-K. Kim, H. Lee, J. Roh, and S.-Y. Lee, “Hierarchical committee of deep cnns with exponentially-weighted decision fusion for static facial expression recognition,” inProceedings of the 2015 ACM on International Conference on Multimodal Interaction. ACM, 2015, pp. 427–434.

[10] F. de la Torre, W.-S. Chu, X. Xiong, F. Vicente, X. Ding, and J. Cohn, “Intraface,” in Automatic Face and Gesture Recognition (FG), 2015 11th IEEE International Conference and Workshops on, vol. 1. IEEE, 2015, pp. 1–8.

[11] Y. Taigman, M. Yang, M. Ranzato, and L. Wolf, “Deepface: Clos-ing the gap to human-level performance in face verification,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2014, pp. 1701–1708.

[12] I. G. Y. Bengio and A. Courville, “Deep learning,” 2016, book in preparation for MIT Press. [Online]. Available: http://www.deeplearningbook.org

[13] “Deep learning,” 2015. [Online]. Available: https://developer. nvidia.com/deep-learning

[14] S. E. Kahou, P. Froumenty, and C. Pal, “Facial expression anal-ysis based on high dimensional binary features,” inECCV 2014. Springer, 2014, pp. 135–147.

[15] S. Moore and R. Bowden, “Local binary patterns for multi-view facial expression recognition,”Computer Vision and Image Under-standing, vol. 115, no. 4, pp. 541–558, 2011.

[16] X. Wang, C. Jin, W. Liu, M. Hu, L. Xu, and F. Ren, “Feature fusion of hog and wld for facial expression recognition,” inSystem Integration (SII), 2013 IEEE/SICE International Symposium on. IEEE, 2013, pp. 227–232.

[17] W. Zheng, H. Tang, Z. Lin, and T. S. Huang, “A novel approach to expression recognition from non-frontal face images,” inComputer Vision, 2009 IEEE 12th International Conference on. IEEE, 2009, pp. 1901–1908.

[18] C. Cortes and V. Vapnik, “Support vector machine,” Machine learning, vol. 20, no. 3, pp. 273–297, 1995.

[19] F. V. Jensen, An introduction to Bayesian networks. UCL press London, 1996, vol. 210.

[20] S. E. Kahou, X. Bouthillier, P. Lamblin, C. Gulcehre, V. Michalski, K. Konda, S. Jean, P. Froumenty, Y. Dauphin, N. Boulanger-Lewandowski et al., “Emonets: Multimodal deep learning ap-proaches for emotion recognition in video,”Journal on Multimodal User Interfaces, pp. 1–13, 2015.

[21] G. Levi and T. Hassner, “Emotion recognition in the wild via convolutional neural networks and mapped binary patterns,” in Proceedings of the 2015 ACM on International Conference on Multi-modal Interaction. ACM, 2015, pp. 503–510.

[22] S. Ebrahimi Kahou, V. Michalski, K. Konda, R. Memisevic, and C. Pal, “Recurrent neural networks for emotion recognition in video,” inProceedings of the 2015 ACM on International Conference on Multimodal Interaction. ACM, 2015, pp. 467–474.

[23] L. Chao, J. Tao, M. Yang, Y. Li, and Z. Wen, “Audio visual emotion recognition with temporal alignment and perception attention,” arXiv preprint arXiv:1603.08321, 2016.

[24] H. Jung, S. Lee, J. Yim, S. Park, and J. Kim, “Joint fine-tuning in deep neural networks for facial expression recognition,” inICCV, December 2015.

[25] M. Valstar and M. Pantic, “Induced disgust, happiness and sur-prise: an addition to the mmi facial expression database,” inProc. 3rd Intern. Workshop on EMOTION (satellite of LREC): Corpora for Research on Emotion and Affect, 2010, p. 65.

[26] G. Zhao, X. Huang, M. Taini, S. Z. Li, and M. Pietik¨ainen, “Fa-cial expression recognition from near-infrared videos,”Image and Vision Computing, vol. 29, no. 9, pp. 607–619, 2011.

[27] M. J. Lyons, S. Akamatsu, M. Kamachi, J. Gyoba, and J. Budynek, “The japanese female facial expression (jaffe) database,” 1998. [28] A. Savran, N. Aly ¨uz, H. Dibeklio ˘glu, O. C¸ eliktutan, B. G ¨okberk,

B. Sankur, and L. Akarun, “Bosphorus database for 3d face anal-ysis,” inBiometrics and Identity Management. Springer, 2008, pp. 47–56.

[29] O. Langner, R. Dotsch, G. Bijlstra, D. H. Wigboldus, S. T. Hawk, and A. van Knippenberg, “Presentation and validation of the radboud faces database,”Cognition and emotion, vol. 24, no. 8, pp. 1377–1388, 2010.

[30] V. G. Kaburlasos, S. E. Papadakis, and G. A. Papakostas, “Lattice computing extension of the fam neural classifier for human facial expression recognition,”IEEE Transactions on Neural Networks and Learning Systems, vol. 24, no. 10, pp. 1526–1538, Oct 2013. [31] G. Recio, A. Schacht, and W. Sommer, “Recognizing dynamic facial

expressions of emotion: Specificity and intensity effects in event-related brain potentials,”Biological psychology, vol. 96, pp. 111–125, 2014.

[32] M. K. Abd El Meguid and M. D. Levine, “Fully automated recog-nition of spontaneous facial expressions in videos using random forest classifiers,”Affective Computing, IEEE Transactions on, vol. 5, no. 2, pp. 141–154, 2014.

[33] T. F. Cootes, G. J. Edwards, and C. J. Taylor, “Active appearance models,” inECCV 1998, 1998.

[34] S. ”Xie and Z. Tu, “Holistically-nested edge detection,” inICCV, 2015.

[35] J. Yang, B. Price, S. Cohen, H. Lee, and M.-H. Yang, “Object contour detection with a fully convolutional encoder-decoder network,”arXiv preprint arXiv:1603.04530, 2016.

[36] H. Noh, S. Hong, and B. Han, “Learning deconvolution network for semantic segmentation,” pp. 1520–1528, 2015.

[37] P. Khorrami, T. Paine, and T. Huang, “Do deep neural networks learn facial action units when doing expression recognition?” in Proceedings of the IEEE International Conference on Computer Vision Workshops, 2015, pp. 19–27.

[38] L. Bottou, “Large-scale machine learning with stochastic gradient descent,” inProceedings of COMPSTAT’2010. Springer, 2010, pp. 177–186.

[39] X. Zhang, L. Yin, J. F. Cohn, S. Canavan, M. Reale, A. Horowitz, and P. Liu, “A high-resolution spontaneous 3d dynamic facial expression database,” inAutomatic Face and Gesture Recognition (FG), 2013 10th IEEE International Conference and Workshops on. IEEE, 2013, pp. 1–6.

[40] S. V. Stehman, “Selecting and interpreting measures of thematic classification accuracy,” Remote sensing of Environment, vol. 62, no. 1, pp. 77–89, 1997.

[41] M. F. Valstar, M. Mehu, B. Jiang, M. Pantic, and K. Scherer, “Meta-analysis of the first facial expression recognition challenge,” Systems, Man, and Cybernetics, Part B: Cybernetics, IEEE Transactions on, vol. 42, no. 4, pp. 966–979, 2012.

[42] P. Viola and M. Jones, “Rapid object detection using a boosted cascade of simple features,” inComputer Vision and Pattern Recogni-tion, 2001. CVPR 2001. Proceedings of the 2001 IEEE Computer Society Conference on, vol. 1. IEEE, 2001, pp. I–511.

[43] Y. Jia, E. Shelhamer, J. Donahue, S. Karayev, J. Long, R. Girshick, S. Guadarrama, and T. Darrell, “Caffe: Convolutional architecture for fast feature embedding,”arXiv preprint arXiv:1408.5093, 2014. [44] A. Krizhevsky, I. Sutskever, and G. E. Hinton, “Imagenet

classifi-cation with deep convolutional neural networks,” inAdvances in neural information processing systems, 2012, pp. 1097–1105. [45] K. Simonyan and A. Zisserman, “Very deep convolutional

networks for large-scale image recognition,” arXiv preprint arXiv:1409.1556, 2014.

(9)

[46] J. Philbin, O. Chum, M. Isard, J. Sivic, and A. Zisserman, “Ob-ject retrieval with large vocabularies and fast spatial matching,” in Computer Vision and Pattern Recognition, 2007. CVPR’07. IEEE Conference on. IEEE, 2007, pp. 1–8.

[47] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,”arXiv preprint arXiv:1512.03385, 2015.

(10)

A

PPENDIX

A

Q

UALITATIVE

R

ESULTS FOR

M

ULTI

-

VIEW

E

MOTION

R

ECOGNITION

Fig. 7.Qualitative Results for Multi-view Emotion Recognition. Top two predictions for each image are displayed, sorted by classification score. Correct categories are shown in green and incorrect categories are shown in red. Observe that, for mis-classified examples, the second prediction of our system is generally correct.

(11)

Fig. 8.Qualitative Results for Multi-view Emotion Recognition. Top two predictions for each image are displayed, sorted by classification score. Correct categories are shown in green and incorrect categories are shown in red. Observe that, for mis-classified examples, the second prediction of our system is generally correct.

(12)

Fig. 9.Qualitative Results for Multi-view Emotion Recognition. Top two predictions for each image are displayed, sorted by classification score. Correct categories are shown in green and incorrect categories are shown in red. Observe that, for mis-classified examples, the second prediction of our system is generally correct.

(13)

Fig. 10.Qualitative Results for Multi-view Emotion Recognition. Top two predictions for each image are displayed, sorted by classification score. Correct categories are shown in green and incorrect categories are shown in red. Observe that, for mis-classified examples, the second prediction of our system is generally correct.

(14)

Fig. 11.Qualitative Results for Multi-view Emotion Recognition. Top two predictions for each image are displayed, sorted by classification score. Correct categories are shown in green and incorrect categories are shown in red. Observe that, for mis-classified examples, the second prediction of our system is generally correct.

(15)

A

PPENDIX

B

Q

UALITATIVE

R

ESULTS

F

OR

A

CTION

U

NIT

D

ETECTION

Inner Brow

Raiser

(AU1)

Outer Brow

Raiser

(AU2)

Brow

Lowerer

(AU4)

Cheek

Raiser

(AU6)

Lid

Tightener

(AU7)

Upper Lip

Raiser

(AU10)

Lip Corner

Puller

(AU12)

Dimpler

(AU14)

Lip Corner

Depressor

(AU15)

Chin

Raiser

(AU17)

Lip

Tightener

(AU23)

True Positives

False Positives

False Negatives

Lip

Pressor

(AU24)

Fig. 12.Qualitative results for Convolutional Action Unit Detection. Percentages refer to the CAUD confidence for predicting the action unit on top of each column. ’True positives’ are defined as positives with confidence above 50%, ’False Positives’ as negatives with confidence above 50%, and ’False Negatives’ as positives with confidence below 50%.

(16)

Inner Brow

Raiser

(AU1)

Outer Brow

Raiser

(AU2)

Brow

Lowerer

(AU4)

Cheek

Raiser

(AU6)

Lid

Tightener

(AU7)

Upper Lip

Raiser

(AU10)

Lip Corner

Puller

(AU12)

Dimpler

(AU14)

Lip Corner

Depressor

(AU15)

Chin

Raiser

(AU17)

Lip

Tightener

(AU23)

True Positives

False Positives

False Negatives

Lip

Pressor

(AU24)

Fig. 13.Qualitative results for Convolutional Action Unit Detection. Percentages refer to the CAUD confidence for predicting the action unit on top of each column. ’True positives’ are defined as positives with confidence above 50%, ’False Positives’ as negatives with confidence above 50%, and ’False Negatives’ as positives with confidence below 50%.

(17)

Inner Brow

Raiser

(AU1)

Outer Brow

Raiser

(AU2)

Brow

Lowerer

(AU4)

Cheek

Raiser

(AU6)

Lid

Tightener

(AU7)

Upper Lip

Raiser

(AU10)

Lip Corner

Puller

(AU12)

Dimpler

(AU14)

Lip Corner

Depressor

(AU15)

Chin

Raiser

(AU17)

Lip

Tightener

(AU23)

True Positives

False Positives

False Negatives

Lip

Pressor

(AU24)

Fig. 14.Qualitative results for Convolutional Action Unit Detection. Percentages refer to the CAUD confidence for predicting the action unit on top of each column. ’True positives’ are defined as positives with confidence above 50%, ’False Positives’ as negatives with confidence above 50%, and ’False Negatives’ as positives with confidence below 50%.

(18)

Inner Brow

Raiser

(AU1)

Outer Brow

Raiser

(AU2)

Brow

Lowerer

(AU4)

Cheek

Raiser

(AU6)

Lid

Tightener

(AU7)

Upper Lip

Raiser

(AU10)

Lip Corner

Puller

(AU12)

Dimpler

(AU14)

Lip Corner

Depressor

(AU15)

Chin

Raiser

(AU17)

Lip

Tightener

(AU23)

True Positives

False Positives

False Negatives

Lip

Pressor

(AU24)

Fig. 15.Qualitative results for Convolutional Action Unit Detection. Percentages refer to the CAUD confidence for predicting the action unit on top of each column. ’True positives’ are defined as positives with confidence above 50%, ’False Positives’ as negatives with confidence above 50%, and ’False Negatives’ as positives with confidence below 50%.

(19)

Inner Brow

Raiser

(AU1)

Outer Brow

Raiser

(AU2)

Brow

Lowerer

(AU4)

Cheek

Raiser

(AU6)

Lid

Tightener

(AU7)

Upper Lip

Raiser

(AU10)

Lip Corner

Puller

(AU12)

Dimpler

(AU14)

Lip Corner

Depressor

(AU15)

Chin

Raiser

(AU17)

Lip

Tightener

(AU23)

True Positives

False Positives

False Negatives

Lip

Pressor

(AU24)

Fig. 16.Qualitative results for Convolutional Action Unit Detection. Percentages refer to the CAUD confidence for predicting the action unit on top of each column. ’True positives’ are defined as positives with confidence above 50%, ’False Positives’ as negatives with confidence above 50%, and ’False Negatives’ as positives with confidence below 50%.

Referencias

Documento similar

In the preparation of this report, the Venice Commission has relied on the comments of its rapporteurs; its recently adopted Report on Respect for Democracy, Human Rights and the Rule

After a deep study of the state of the art for the task of person detection, it has been possible to train a neural network based on the Yolo v2 detection, using MobileNet and Tiny

Chapter 6 introduces a novel approach to Voice Activity Detection (VAD) on noisy environments capable of processing the raw speech signal using a combination of Convolutional

Finally, we adapt also the detection proposal to deal with specific problems of object recognition in different areas. On the one hand, we present a new fully automatic computer

Many methods have been proposed for abandoned and stolen object detection focusing on the stabilization of the image sequence from a moving camera [2], based on the

Building on recent results, we write the action of the angular momentum operators on scattering amplitudes as derivatives with respect to the Mandelstam invariants to uncover a

In this work, we propose an end-to-end approach to the language identification (LID) problem based on Convolutional Deep Neural Networks (CDNNs).. The use of CDNNs is mainly motivated

Another powerful model for MS sample fingerprinting are convolutional neural networks (CNN), an image recognition model recently applied in automated targeted analysis of