• multi-stream data:
Since a large number of markers are placed on the
face, it is necessary to process the multiple streams
by considering the interrelation between them.
These data have information about the position of
each part of the face with good accuracy.
Since these data obtained by motion capture system
have just the positions of the markers, they need to be
transformed to parameters representing the features of
each facial expression notably. We present a method
for extracting essential parameters that cause visible
change of the face by estimating the force vector of
each point on the face. The force vectors need to be
calculated from displacements of points on the face.
That is, the forces acting on the skin of the face are
obtained by the movments of each point on the face.
Therefore, our approach uses a method of analyzing
inverse problem using FEM (Cook, 1995) to estimate
the force vectors. FEM is a very powerful tool to ob-
tain stress and strain when outside forces act on an
object. Since FEM can be easily applied to various
engineering problems and handles complex loading,
the method for analyzing inverse problem using FEM
is the best tool as a method to estimate the force vec-
tors from the displacemants.
Then, the force vectors must be compared correctly
to recognize facial expressions. The comparison of
force vectors by Euclidean distance does not have
essential significance, since the force vectors have
two elements: direction and length. Thus, we pro-
pose a new similarity metric AMSS (Angular Met-
rics for Shape Similarity) for an effective evaluation
of force parameters. A similarity of force vectors by
AMSS is calculated from the difference of the length
and the angle between the two vectors. The com-
parison method using AMSS can achieve an exact
evaluation using elements of vectors effectively. Ex-
pression recognition is done by estimating force vec-
tor in DTW (Dynamic Time Warping) (Sankoff and
Kruskal, 1983) using AMSS.
This paper is organized as follows: Section 2 de-
scribes the motion capture system and the data for
facial expression. Section 3 and Section 4 describe
the technique for extracting parameters representing
forces acting on points on a face using FEM, and
the method of facial expression recognition using the
force vectors. Section 5 performs an experiment on
expression recognition and discusses the results. We
conclude in Section 6 by summarising the paper and
suggesting future research directions.
2 FACS
In this paper, expressions are recognized based on the
idea of FACS which is widely used in the field of fa-
cial expression analysis. Since a facial expression is
indeed the combination of movements of the points
on the face, the data of the whole surface of the face
is not needed for the recognition of facial expressions.
All you need to recognize the facial expressions relies
on the information of some points on a face. We also
use the optical motion capture system to obtain the
data of 35 points on the face.
FACS is a method for measuring and describing fa-
cial behaviors proposed by Ekman and Friesen. FACS
is widely used in the field of the research about the
recognition of facial expression (Essa and Pentland,
1997) (Lien et al., 1998). Ekman and Friesen devel-
oped the original FACS by determining how the con-
traction of each facial muscle (singly and in combi-
nation with other muscles) changes the appearance of
the face.
In FACS, an expression is described by a basic
unit called AU. AU is a primitive unit of the expres-
sion movement that can be identified visually, and
there are 44 kinds of AUs in total. Human expres-
sions are described by these combinations. For exam-
ple, Sadness is described as “1+4+15+23” since it is
composed of four AUs, such as 1, 4, 15 and 23.
The 17 AUs used to express basic facial expres-
sions are shown in Table 1. For instance, AU 1 de-
scribes the movement of raising the inner corner of
the eyebrow and AU 4 describes the movement of
puckering up one’s brows. Each AU is influenced by
a specific expression muscle respectively.
Table 1: Examples of AU(Action Unit).
No. Name No. Name
1 InnerBrowRaise 14 Dimpler
2 OuterBrowRaise 15 LipCornerDepress
4 BrowLower 16 LowerLipDepress
5 UpperLidRaise 17 ChinRaise
6 CheekRaise 20 LipStretch
7 LidT ight 23 LipT ight
9 NoseW rinkle 25 LipsP art
10 UpperLipRaise 26 JawDrop
12 LipConerP ull
Six basic expressions proposed by Ekman are
widely used in classification of a human expression.
The six basic expressions are as follows: happiness,
sadness, surprise, disgust, fear and anger. Table 2
shows the combination and the strength of AUs to rep-
resent the six basic expressions. The numerical value
in parentheses describes strength. It is 0 when the AU
is invisible, and it is 100 when the AU is apparent. For
instance, Anger is composed of eight AUs, such as 2,
4, 7, 9, 10, 12, 15 and 26, because features of Anger
are puckering up one’s brows, staring, applying, and
clenching teeth.
In this paper, expression data are obtained by using
an optical motion capture system. These data consist
of x, y, z coordinates for each point on the testee’s
FACIAL EXPRESSION RECOGNITION BASED ON FACIAL MUSCLES BEHAVIOR ESTIMATION
49