Table 6: Lower action unit recognition results on hemifaces
using optical flow.
N. RR RR (LH) RR (RH)
4 86.81% 89.83% 88.14%
6 78.72% 82.33% 79.51%
8 70.06% 72.0% 68.62%
10 67.44% 67.24% 64.66%
12 66.21% 65.76% 63.32%
note:N.= number of classes; RR=recognition rate; LH=left
hemiface; RH=right hemiface.
Consequently, for the global feature extraction,
Gabor wavelet feature extraction can reach better
recognition rate than optical flow feature extraction;
for the local feature extraction, optical flow can reach
better recognition rate in the above eyes region. We
combined both of these feature extraction methods to
achieve best recognition result.
To get accurate recognition result, we first rec-
ognize the action units using the classifier of most
classes, if the recognized action units combination is
included in other classifiers, these classifiers will also
be used to verify the recognition result.
6.2 Facial Action Unit Inference
Using the AU codes of 452 samples of facial expres-
sions as the training data, we learned the BN of 22
action units, as seen in figure 2. Compared to the
BN of 14 action units learned by (Tong et al., 2007),
more nodes and links are learned in our research.
This means that there are more complex relationships
among the AUs. Based on the recognized AUs as ev-
idence, the DBN can infer related AUs according to
the corresponding predicted probabilities.
Figure 2: The learned BN of 22 action units.
Specially, the AUs we don’t recognize directly
because of lack of samples such as AU13(Sharp
lip puller), AU16(Lower lip depress), AU24(Lip
presser), AU26(Jaw drop) and AU31(Jaw clencher)
can be inferred with their probabilities. For ex-
ample, according to the CPTs of the learned DBN,
P(AU31 = 1|AU26 = 1) = 0.7308, means that when
Jaw drop occurs, Jaw clencher will also occur with
the probability of 0.7308.
7 CONCLUSIONS
Aiming to recognize facial action units efficiently, we
analyze the Gabor wavelet and optical flow feature ex-
traction in global and local facial regions, and use sup-
port vector machine and dynamic bayesian network
for classification and inference respectively. The pro-
posed method is capable of recognizing and inferring
most action units in FACS, and can reach good per-
formance.
ACKNOWLEDGEMENTS
This work is supported by National Nature Science
Foundation of China (No.60873269, No.61103097),
International Cooperation between China and Japan
(No.2010DFA11990) and the Fundamental Research
Funds for the Central Universities.
REFERENCES
Kanade, T., Cohn, J., and Tian, Y. L. (2000). Compre-
hensive database for facial expression analysis. In
Proc. of 4th IEEE Int. Conf. on Automatic Face and
Gesture Recognition, pages 46–53, Washington, DC,
USA. IEEE Computer Society.
Kapoor, A., Qi, Y., and Picard, R. W. (2003). Fully auto-
matic upper facial action recognition. In Proc. of IEEE
Int. Workshop on Analysis and Modeling of Faces and
Gestures, pages 195–202.
Korb, K. B. and Nicholson, A. E. (2004). Bayesian Artificial
Intelligence. A CRC Press Company.
Pantic, M. and Rothkrantz, L. J. M. (2000). Expert system
for automatic analysis of facial expressions. Image
and Vision Computing, 18(11):881–905.
Tian, Y. L., Kanade, T., and Cohn, J. F. (2001). Recogniz-
ing action units for facial expression analysis. IEEE
Trans. on PAMI, 23(2):97–115.
Tong, Y., Chen, J. X., and Ji, Q. (2010). A unified
probabilistic framework for spontaneous facial action
modeling and understanding. IEEE Trans. on PAMI,
32(2):258–273.
Tong, Y., Liao, W. H., and Ji, Q. (2007). Facial action unit
recognition by exploiting their dynamic and semantic
relationships. IEEE Trans. on PAMI, 29(10):1–17.
Viola, P. and Jones, M. (2001). Rapid object detection us-
ing a boosted cascade of simple features. In Proc. of
IEEE Conf. on CVPR, pages 511–518, Washington,
DC, USA. IEEE Computer Society.
FACIAL ACTION UNIT RECOGNITION AND INFERENCE FOR FACIAL EXPRESSION ANALYSIS
697