In this work, we present a robust and lightweight approach for the automatic
fitting of deformable 3D face models to facial pictures. Well known fitting methods,
for example those taking into account statistical models of shape and appearance, need
a training stage based on a set of facial landmarks, manually tagged on facial pictures.
In this manner, new pictures in which to fit the model cannot differ excessively in
shape and appearance (including illumination changes, facial hair, wrinkles, and so on)
from those utilized for training. By contrast, our methodology can fit a generic face
model in two stages: (1) the localization of facial features based on local image
gradient analysis; and (2) the backprojection of a deformable 3D face model through
the optimization of its deformation parameters. The proposed methodology preserves
the advantages of both learning-free and learning-based methodologies. Subsequently,
we can estimate the position, orientation, shape and actions of faces, and initialize userspecific
face tracking approaches, such as Online Appearance Models (OAMs), which
have demonstrated to be more robust than generic user tracking methodologies.
Experimental results demonstrate that our strategy outperforms other fitting methods
under challenging illumination conditions and with a computational footprint that
permits its execution in gadgets with reduced computational power, such as cell phones
and tablets. Our proposed methodology fits well with numerous systems addressing
semantic inference in face images and videos.
Keywords: 2D shape landmarks, 3D face model, Deformable model back projection,
facial actions, Facial expression recognition, Facial feature extraction,
facial parts, Face gesture analysis, Face model fitting, Face recognition, Face
tracking, Gradient maps, Head pose estimation, Learning-free, Levenberg-
Marquardt, Monocular image, Online appearance model, Shape variations,
Sigmoidal filter, Weak perspective.