Professional Documents
Culture Documents
Abstract—In this paper we investigate the performance are based on higher-order statistics (HOS), however the dimen-
of different feature extraction methods for facial expression sionality of the resulting data is so high. The dimensionality
recognition based on the higher-order local autocorrelation reduction can be achieved by selection more informative
(HLAC) coefficients and Gabor wavelet filters. We use a
Cohn-Kanade database of facial images, organized in training features based on mutual information.
and testing sets, for evaluation. Autocorrelation coefficients The mutual information (MI) [5] parameter was investigated
are computationally inexpensive, inherently shift-invariant and as a measure of information overlap between features. The
quite robust against changes in facial expression. The focus
is on the difficult problem of recognizing an expression in
mutual information was used as an objective criterion in
different resolutions. Results indicate that local autocorrelation selection of optimal sub-sets of features in a feature reduction
coefficients have surprisingly high information content. task.
In contrast to the classical correlation-based feature selec-
-Index Terms- Feature extraction, higher-order statistics, facial
expression recognition, Gabor filters. tion methods, the mutual information can measure arbitrary
relations between variables and it does not depend on trans-
formations applied to different variables. It can be potentially
I. I NTRODUCTION useful in problems where methods based on linear relations
between data are not performing well.
F ACIAL expression is a visible manifestation of the af-
fective state, cognitive activity, intention, personality, and
psychopathology of a person; it not only expresses our emo-
A functional block-diagram of the proposed facial expres-
sion recognition system is illustrated in Fig.1.
tions, but also provides important communicative cues during
social interaction. Reported by psychologists, facial expression
constitutes 55% of the effect of a communicated message
while language and voice constitute 7% and 38% respectively.
So it is obvious that analysis and automatic recognition of
facial expression can improve human-computer interaction or
even social interaction.
An automatic classification of facial expressions consists
of two stages: feature extraction and feature classification.
The feature extraction is a key importance to the whole
classification process. If inadequate features are used, even
the best classifier could fail to achieve accurate recognition.
In most cases of facial expression classification, the process
of feature extraction yields a prohibitively large number of
features and subsequently a smaller sub-set of features needs
to be selected according to some optimality criteria.
Lyons et al. [2] adopted a wavelet-based face representation.
Input images were convolved with the Gabor filters of five Fig. 1: Block diagram of the facial recognition system
spatial frequencies and the amplitude of the complex-valued
filter responses were sampled on 34 manually selected facial
points and combined into a single vector, containing 1020 The remainder of this paper provides detailed descriptions
elements. Zhang et al. [3] used a similar representation while of the proposed system, experiments and results. Section 2
they applied wavelet of 3 scales and 6 orientations. They explains the image pre-processing stages. In Section 3, the
also considered geometric position of the 34 facial points as feature extraction methods are explained. Section 4 describes
features. the feature selection. In Section 5, the classification process
In this study Higher-Order Local Auto-Correlation (HLAC) based on the NB classifier is explained. Section 6 contains the
features are used for Feature Extraction, HLAC features, an experiments and results, and Section 7 presents the conclu-
extension of autocorrelation features (second-order statistics), sions.
Fig. 2: Six facial expression images after pre-processing. Fig. 3: 25 mask patterns of the HLAC features (3x3).
facial expressions. The tested images were classified using TABLE III: Confusion table for Gabor features (person-
HLAC features and NB classifier. Independent).
For comparison, other popular method, Gabor features [12], Anger Disgust Fear Happy Sad Surprise
was applied to the same classification problems. We used Anger 63.3 20.0 6.7 0 10.0 0
Disgust 30.3 48.5 3.0 6.1 12.1 0
filter banks designed in [12]. The numbers of scales (5) and Fear 11.5 9.0 55.1 16.7 2.6 5.1
orientations (8) of the filter bank were adjusted in feature Happy 9.2 1.0 18.5 71.3 0 0
extraction. Sad 20.8 9.2 3.4 0 58.3 8.3
Surprise 3.6 3.5 3.0 0 8.9 81.0
The classification results are presented in Fig.7 and Fig.8.
The results are shown for different resolutions from 16x16
to 128x128. Table.II and Table.III shows the confusions for
different expression in low resolution(16x16) . Fig.6 illustrated VII. C ONCLUSION
the average correct classification for both HLAC and Gabor A comparison of feature extraction methods for the facial
features. It is shown that the classification result based on expression recognition from image sequences was presented
HLAC features for low resolution samples are better than and tested. The method is fully automatic and includes:
Gabor features,however, for high resolution samples the classi- face detection, maximum arousal detection, feature extraction,
fication performance for Gabor features is more accurate than selection of optimal features and classification. The higher-
the HLAC features. order local autocorrelation (HLAC) features in low resolution
Table.I shows an example of the CPU times corresponding images increased the average percentage of correct classifi-
to different resolutions of the feature extraction process. The cations from 62.9% to 65.9% and from 65.2% to 67.2% for
times are given (CPU speed 2.4 GHz and 2GB RAM) for two 16 × 16 and 32 × 32 respectively. Furthermore, the CPU time
cases: the feature extraction process based on Higher order decreased from 2.02s to 0.22s and 2.25s to 0.29s for 16 × 16
local autocorrelation and the feature extraction process based and 32 × 32 respectively.
on Gabor filters. The presented feature selection method is based on the
mutual information(MI) criterion, and does not assume linear
TABLE I: Comparison of feature extraction time.
dependencies between data. It can therefore handle arbitrary
HLAC Features Gabor Features relations between the pattern coordinates and the different
Image Resolution Time(sec) Image Resolution Time(sec)
16x16 0.22 16x16 2.02
classes. The additional advantages of the feature selection
32x32 0.29 32x32 2.25 based on the MI criterion include computational simplicity and
64x64 0.27 64x64 3.57 invariance to the data transformations. The system not only
128x128 0.50 128x128 9.65
offers an optimized feature selection, but also automatically
finds an optimal frame to represent a given class of emotion.
An overall improvement of the classification results and
the discrimination between different facial expressions was
observed when using HLAC features. The accuracy for high
resolution images based on Gabor filters were better than
HLAC features, however the complexity and time consuming
were more discriminate in Gabor filter feature extraction
process.
R EFERENCES
[1] Kanade, T., Cohn, J. F., and Tian, Y..”Comprehensive database for facial
expression analysis,”Proceedings of the Fourth IEEE International Con-
ference on Automatic Face and Gesture Recognition, Grenoble, France,
pp. 46-53, 2000.
[2] M. Lyons, J. Budynek, and S. Akamastu, ”Automatic Classification
of Single Facial Images”,IEEE Trans. Pattern Analysis and Machine
Intelligence, Vol.21, 1999, pp. 1357- 1362.
[3] Z. Zhang, M. Lyons, M. Schuster, and S. Akamatsu, ”Comparison
Between Geometrybased and Garbor-Wavelet-based Facial Expression
Recognition Using Multi-layer Perceptron”,Proc. 3rd Int. Conf. Automatic
Face and Gesture Recognition, 1998, pp. 454-459.
[4] N. Otsu and T. Kurita. ”A new scheme for practical flexible and intelligent
vision systems”. In Proceedings of the IAPR Workshop on Computer
Vision, pp. 431435, 1988.
[5] Battiti R., ”Using Mutual Information for Selecting Features in Super-
vised Neural Net Learning”. IEEE Trans on Neural Networks, vol.5, no.
4, July 1994, pp. 537-550..
[6] Viola P., Jones M., ”Robust Real-time Object Detection,”, International
Journal of Computer Vision,2004.
[7] Kwak N., Choi C., ”Input Feature Selection for Classification Problems,”,
IEEE Trans. On Neural Networks, vol.13, no.1, pp.143-159, 2002.
[8] T. Toyoda and O. Hasegawa. ”Texture classification using extended higher
order local autocorrelation features,” Proceedings of the 4th International
Workshop on Texture Analysis and Synthesis, pp. 131136, 2005.
[9] Duda R.O., P. E. Hart P.E., Stork D.G., ”Pattern Classification.” New
York: John Wiley and Sons. Inc., 2001.
[10] Rish I., ”An empirical study of the naive Bayes classifier” IJCAI
Workshop on Empirical Methods in Artificial Intelligence, 2001.
[11] Liu, F., et al., ”Facial expression recognition using HLAC features and
WPCA,” Springer Verlag, Heidelberg, Beijing, China, pp.88-94, 2005.
[12] Lajevardi S.M., Lech M., ”Averaged Gabor Filter Features for Facial
Expression Recognition”, DICTA08, Canberra, Australia, 2008.
[13] Lajevardi S.M., Lech M., ”Facial Expression Recognition Using a Bank
of Neural Networks and Logarithmic Gabor Filters”, DICTA08, Canberra,
Australia, 2008.
[14] Lajevardi S.M., Lech M., ”Facial Expression Recognition from Image
Sequences Using Optimised Feature Selection”, IVCNZ08, Christchurch,
New Zealand, 2008.