January 10, 2010 / Vol. 8, No. 1 / CHINESE OPTICS LETTERS 29
Exaggeration of facial expressions from facial motion
capture data
Seongah Chin (ªªª)
∗
and Chung-Yeon Lee (ooo¿¿¿,,,)
Division of Multimedia, College of Engineering, Sungkyul University,
#147-2 Anyang-8 Dong. Manan-Gu, Anyang City, 430-742, Korea
∗
E-mail: solideo@sungkyul.edu
Received June 17, 2009
We propose a method for exaggerating facial expressions derived from exaggeration mappings that trans-
form facial motions into exaggerated motions. The exaggeration mapping of facial motions is defined
by non-negative matrix factorization. Three-dimensional facial expressions are simulated using the exag-
gerating rate as an input value to convey the degree of variation. Experiments show the validity of the
exaggeration mapping and facial expression simulations.
OCIS co des: 100.0100, 110.0110.
doi: 10.3788/COL20100801.0029.
Facial expressions have been widely used for characters,
digital actors, and virtual humans in various fields such as
video games, movies, social agents, commercials, and an-
thropology. The human face, in particular, plays a signif-
icant role in non-verbal communication, with a plethora
of expressions that vary by personality. Most researchers
in the area of facial expression synthesis seem to focus on
realistic expressions or retargeting
[1−3]
, rather than exag-
gerating expressions. To draw attention to subjects, it is
very important to exaggerate facial expressions, but this
is a labor-intensive process. As a consequence, a method
for automatic exaggerated facial expression synthesis is
a must. Exaggeration is considered to be similar to plac-
ing more emphasis on the key features of an object. It
should be noted that the larger accentuated features are
found before the lesser features are determined. However
to our knowledge, no published articles have reported on
the exaggeration of facial expressions. In this letter, we
propose a method to exaggerate facial expressions de-
rived from facial motions acquired by motion capture. In
this transformation, we employ non-negative matrix fac-
torization (NMF) to realize exaggeration mapping. The
proposed method is validated by showing some experi-
mental results.
Exaggeration mapping is used to transform a facial mo-
tion composed of feature points on a face acquired by
motion capture into an exaggerated facial motion. The
fundamental idea of this approach to the exaggeration
mapping of facial motions is to place more emphasis on
those features with relatively more movements. The first
step is proper data factorization. NMF is suitable for
parts-based data factorization because of its advantages
over principal component analysis (PCA) or vector quan-
tization (VQ) with respect to its use of non-negative con-
straints. In addition, NMF learns parts of faces, whereas
PCA and VQ learn holistically. These constraints lead to
a parts-based representation computed only by additive,
rather than subtractive, combinations
[4,5]
.
The mapping method primarily comprises two proce-
dures. Firstly, NMF decomposes a facial motion matrix
into a basis and its weight matrix. Then, a facial motion
is exaggerated by multiplying both the weights and resid-
uals acquired from the NMF decomposition by a specific
exaggeration rate.
Cross-cultural research presents six universal expres-
sions — surprise, fear, disgust, anger, happiness, and sad-
ness — for which a sequence of facial markers, called a fa-
cial motion, is factorized in order to exaggerate individual
facial markers of the facial motion. Each motion consists
of the three-dimensional (3D) movements of markers at-
tached to the principal facial muscles of an actor. These
facial motions are regarded as an n × m matrix M, each
column of which consists of the x, y, and z coordinates of
the markers. Given a facial motion denoted by M, NMF
decomposes M into two matrices B and E, as
M
iµ
≈ (BE)
iµ
=
X
r
a=1
B
ia
E
aµ
, (1)
to approximate the facial motions. The dimensions of
the factorized matrices B and E are n × r and r × m,
respectively, with r satisfying (n + m)r < nm.
Each column of the matrix B contains a basis vector,
while each column of E includes the weights correspond-
ing to the measurement column in M using the basis from
B. To estimate the factorization matrices, an objective
function has to be defined. This objective function works
out the likelihood of computing the facial motions in M
from the basis B and encodings E. The objective func-
tion that we used is given as
H =
X
n
i=1
X
m
µ=1
[M
iµ
log(BE)
iµ
− (BE)
iµ
]. (2)
Solutions for NMF begin with initializing the non-
negative conditions for B and E. Continuing the iter-
ation of the update rules in
B
ia
← B
ia
X
µ
M
iµ
(BE)
iµ
E
aµ
, (3)
E
aµ
← E
aµ
X
i
B
ia
M
iµ
(BE)
iµ
, (4)
M finds an approximate factorization M ≈ BE by con-
verging to a local maximum of the objective function
1671-7694/2010/010029-04
c
° 2010 Chinese Optics Letters