Conference PaperPDF Available

Face Gender Classification on Consumer Images in a Multiethnic Environment

Authors:

Abstract and Figures

In this paper, we target at face gender classification on consumer im- ages in a multiethnic environment. The consumer images are much more chal- lenging, since the faces captured in the real situation vary in pose, illumination and expression in a much larger extent than that captured in the constrained en- vironments such as the case of snapshot images. To overcome the non- uniformity, a robust Active Shape Model (ASM) is used for face texture nor- malization. The probabilistic boosting tree approach is presented which achieves a more accurate classification boundary on consumer images. Besides that, we also take into consideration the ethnic factor in gender classification and prove that ethnicity specific gender classifiers could remarkably improve the gender classification accuracy in a multiethnic environment. Experiments show that our methods achieve better accuracy and robustness on consumer im- ages in a multiethnic environment.
Content may be subject to copyright.
Face Gender Classification on Consumer Images in a
Multiethnic Environment
Wei Gao and Haizhou Ai
Computer Science and Technology Department,
Tsinghua University, Beijing 100084, China
ahz@mail.tsinghua.edu.cn
Abstract. In this paper, we target at face gender classification on consumer im-
ages in a multiethnic environment. The consumer images are much more chal-
lenging, since the faces captured in the real situation vary in pose, illumination
and expression in a much larger extent than that captured in the constrained en-
vironments such as the case of snapshot images. To overcome the non-
uniformity, a robust Active Shape Model (ASM) is used for face texture nor-
malization. The probabilistic boosting tree approach is presented which
achieves a more accurate classification boundary on consumer images. Besides
that, we also take into consideration the ethnic factor in gender classification
and prove that ethnicity specific gender classifiers could remarkably improve
the gender classification accuracy in a multiethnic environment. Experiments
show that our methods achieve better accuracy and robustness on consumer im-
ages in a multiethnic environment.
Keywords: Boosting tree, gender classification, multiethnic environment
1 Introduction
Face vision research has achieved significant advancement in the past decade espe-
cially on face detection and face alignment or facial feature location technologies that
can readily provide effective tools to extract faces from raw images. With faces hav-
ing being extracted, demography classification that includes gender, ethnicity and age
become an interesting topic due to its potential applications in photo album manage-
ment, shopping statistics for marketing, visual surveillance, etc. Unlike ethnicity and
age estimation, gender classification has attracted more attention in face classification
literature since it is the most basic information from a face which human can have a
very clear division in perception.
In the early days, most of researches in gender classification are about human’s
perceiving for gender from a psychology point of view, where the computer is used
just as an assistant tool and no automatic gender classification system is developed.
More recently neural network methods were firstly used in gender classification.
Golomb et al. [1] trained a gender classifier “SexNet” with a two-layer neural net-
work on 90 facial images and achieved an accuracy of 91.9%. Gutta et al. [2] trained a
neural network on 3000 faces from FERET dataset and decreased the error rate to 4%.
Balci [3] used eigenfaces and trained a Muti-Layer Perceptron on FERET Dataset to
analyze which eigenface contributed to gender classification. Later, Moghaddam et al.
[4] used SVM and achieved an accuracy of 96.6% on FERET’s 1755 faces which was
the best result on this set. However for human, only about 96% in accuracy can be
achieved using face information only on this problem according to [5]. BenAbdlkader
et al. [6] extracted both local and holistic features and used LDA and SVM on a data-
set of about 13,000 faces that achieved 94.2% correct rate. Yang et al. [7] used LDA,
SVM and Real AdaBoost respectively on 11,500 snapshots that achieved about 96.5%
correct rate. Another interesting work, Lapedriza et al. [9] analyzed the external face’s
contribution to gender classification and developed a classifier [10] based on both
internal and external faces that resulted in 91.7% correct rate on FRGC dataset.
In overview of most of previous works on gender classification, one thing in com-
mon is that all those face images used in the experiments are caught in constraint
environments, and further they are using 5-fold CV verification method to evaluate
performance that implies their test sets have the same distributions as their training
sets. So the generalization ability on independent sets is still a problem.
Shakhnarovich et al. [8] reported that gender classification on images from internet by
AdaBoost algorithm can only achieve 78% accuracy and by SVM can only achieve
77% accuracy. Besides, the ethnic factor is less considered before. The gender classi-
fier trained can not guarantee good generalization ability in a multiethnic environment.
In this paper we target at face gender classification on consumer images of which
faces vary greatly in pose, illumination and expression. Active Shape Model (ASM) is
used for normalization and a boosting tree is trained on about 10,100 face images
from unconstrained environment. Comparative experiments with other methods in-
cluding SVM and Real AdaBoost on independent test sets are reported to show its
effectiveness. To handle a multiethnic environment, we treat face’s ethnicity as a
latent variable and the ethnicity specific gender classifiers are trained.
The rest of this paper is organized as follows: Section 2 gives an overview of our
gender classification system, Section 3 describes the boosting tree classification
method, Section 4 gives the gender classifier structure in multiethnic environment,
and finally, Section 5 is the conclusion.
2 Gender classification on consumer images
By consumer images we mean those digital photo images caught by popular users of
digital cameras. Compared with faces caught from constraint environment, such as
snapshots, faces in consumer images are more diverse in resolution, makeup, as well
as in illumination, pose and expression (as shown in Fig.1), therefore they are more
challenging to deal with in classification. In this situation, preprocessing and normali-
zation become a critical issue.
As for gender classification methods, AdaBoost has been proved very effective in
both accuracy and speed in the literature. Since AdaBoost is much faster than SVM,
for potential practical applications we choose it to develop a boosting based method
for gender classification. In fact, AdaBoost can mine discriminative features auto-
matically from a large set by giving miss-classified samples more attention. Yang et
al. [7] and Lapedriza et al. [9][10] showed that boosting algorithm achieved compara-
tive accuracy with SVM in gender classification problem. But the main drawback of
this algorithm is overfitting after over a certain number of iterations which means
poor generalization ability in other dataset, especially on those with high intra-class
variations. Since faces in consumer images are with great intra-class variations, it is
found very difficult to learn a single boosting classifier as in [7] [9] [10] in our ex-
periments, therefore divide and conquer strategy becomes necessary for better per-
formance.
Fig.1. Faces from Consumer Images
For a flowchart of our gender classification system, see Fig.2 First a face detection
algorithm [12] is used to detect faces from consumer images and then a variation of
ASM method [13] is used to locate 88 facial feature points for each detected face. For
normalization, a shape free face texture is acquired by triangular warping from a
shape aligned face to the32 32×mean face shape. Compared with the conventional
eye-center normalization in face recognition approaches, this method eliminates some
pose and expression variations.
Fig.2. Face Gender Classification Flowchart
3 Gender classification by boosting tree
3.1 Probabilistic boosting tree
The probabilistic boosting tree (PBT) method is originally proposed by Tu [11] to
deal with the problem of object categorization in natural scenes. It is a new divide-
and-conquer strategy with soft probability boundary. The boosting tree method find
the classification boundary step wisely by putting the ambiguous samples to both left
and right sub-trees as shown in Fig.3 (left). Gradually, more similar samples will be
sent to sub-tree nodes which results in a reduction of intra-variation. The boosting tree
can approach the target posterior distribution by tree expansion.
In the boosting tree structure, each node is a strong classifier trained by AdaBoost
algorithm. We adopt the LUT-based Real AdaBoost method in [14] and use simple
Haar-like features [16] to construct weak classifiers. After T iterations of learning, the
strong classifier has the form:
1
() ()
T
tt
t
Hh
α
=
=
xx
where ()
t
hxis the t-th weak classifier, t
α
is the normalize coefficient and ()H
x
is
the output confidence.
To construct a boosting tree, the confidence output by the root node is further
mapped to probability by sigmoid function as proposed in [15]:
()
exp{2 ( )}
1| 1exp{2()}
Hx
qx
H
x
+=
+,
()
exp{ 2 ( )}
1| 1 exp{ 2 ( )}
Hx
qx
H
x
−=
+−
where
(
)
1|qx+,
()
1|qxdonates the sample’s probability to be positive or to be
negative respectively. Based on the probability above, we split the training set into
sub-trees. This is done by choosing a threshold parameter
ε
to divided probability into
three intervals as shown in Fig.3 (left), that is, the left tree with samples in 1
[0, )
2
,
the right tree with samples in 1
(,1]
2
ε
+ and the ambiguous samples in
11
[,]
22
ε
ε
−+
will be added into both the left and the right sub-tree (as show in Fig.3
(right)). In practice, instead of using a fixed threshold for every tree nodes as in [11],
we choose a variable threshold for each node according to the distribution of samples
to make the tree trained more balanced.
00.1 0.2 0. 3 0.4 0.5 0.6 0. 7 0.8 0. 9 1
0
20
40
60
80
100
120
140
160
180
1
2
ε
+
1
2
ε
To left
sub-tree
To right
sub- tree
ambiguous
samples
probability
Number of
smaples
Fig.3. (left) Histogram of probability distribution of positive and negative samples and three
intervals divided. (right) Probabilistic boosting tree structure. Faces in the left and right of the
nodes correspond to positive and negative samples.
The above procedure is repeated to construct a tree of which each node is a confi-
dence-rated strong classifier learned by Real AdaBoost algorithm.
With the PBT trained, given a face sample, its normalized face is fed into the root
node of the tree to start the decision procedure iteratively. At each node the probabil-
ity to be a positive sample and that to be a negative sample are denoted as
()
1|
P
qx+ and
()
1|
N
qx respectively. And then it will be fed into both its left
and right sub-tree to compute its corresponding probabilities ()
right
p
yand
()
left
p
y. The final decision is computed as:
( | ) ( 1| ) ( ) ( 1| ) ( )
right left
p
yxqxpyqxpy=+ +−
3.2 Experiment result
Experiments are carried out on two kinds of face image datasets: snapshot datasets,
and consumer image datasets. And for each kind of face images, two face sets are
established: one for training and 5-fold CV test, while the other is totally independent
from the training set which is used to judge the algorithms’ generalization ability.
The snapshot training dataset (Snapshots) consists of 15,300 faces from controlled
environment. The snapshot faces are all frontal with similar lighting condition and
expressions. And the independent snapshot set consists of 1800 faces (Independent
Snapshots). The consumer image training dataset consists of about 10,100 Mongoloid
faces in real environment (Consumer Images) with significant changes in poses, illu-
mination and expressions. Similarly another independent consumer image dataset is
collected which consists of 1,300 faces (Consumer Images). All the face datasets
collected above contain nearly equal number of samples for each gender. The boost-
ing tree method is compared with SVM and Real AdaBoost on those datasets.
Table 1 gives both the results on the Snapshot dataset under the 5-fold CV verifica-
tion protocol and the results tested on two other datasets of which ‘All consumer
images’ means the sum of the two consumer image datasets. The SVM method uses
Gaussian kernel. The Real AdaBoost method uses a strong classifier learned after 500
rounds. The boosting tree is composed of 15 strong classifier nodes and its depth is 4.
The generalization ability is evaluated on the independent snapshots and consumer
image dataset.
Table.1. Results on snapshot dataset under 5-fold CV and results tested on two independent
datasets. 5-fold
CV Independent
Snapshots All consumer
images
SVM 96.38 87.89 66.37
AdaBoost 96.50 90.41 80.50
PBT 97.13 93.48 82.07
Table.2. Results on Consumer Images under 5-fold CV and results tested on two independ-
ent datasets. 5-fold
CV Independent
Consumer
Images
All
snapshots
SVM 90.24 88.13 90.72
AdaBoost 94.12 88.61 92.89
PBT 95.51 92.84 93.71
Table 2 gives both the results on the Consumer Image dataset and the results tested
on two other datasets of which ‘All snapshots’ means the sum of the two snapshot
image datasets. As before, the Real AdaBoost method uses a strong classifier learned
after 500 rounds, and the PBT with 15 nodes and a depth of 4.
From the above results, we can see from Table 1 that all the three methods
achieved comparative performance in snapshot datasets while their generalization
ability on the consumer images is bad. However the PBT achieved better generaliza-
tion ability than the other two methods on independent snapshot dataset. From the
Table 2, we can see on the consumer images, PBT’s generalization ability remarkably
outperforms SVM and Real AdaBoost, and their generalization ability on the snapshot
dataset is comparative with the classifier directly trained on snapshots. So, although
there are variations between indoor controlled environments and unconstraint envi-
ronments, the classifier trained on real consumer images from unconstraint environ-
ments can achieve better generalization ability. We can conclude that the PBT method
can describe the classification boundary more accurately than the other two.
4 Gender classification in a multiethnic environment
Compared with gender classification, ethnicity classification attracts less attention in
demography classification. Intuitively ethnicity classification could be done almost in
the same way as gender classification technically. But different from gender classifi-
cation, ethnicity classification is much harder and sometimes even human can not
have a very clear division for ethnicity in perception. In literature, G. Shakhnarovich
et al. [8] divided ethnicity into two categories: Asian and Non-Asian, while in [7] [18]
[19] three categories with Mongoloid, Caucasoid and African were adopted, and in
[17] four ethnic labels with Caucasian, South Asian, East Asian, and African are used.
In this paper, we use three ethnic labels with Mongoloid, Caucasian and African.
4.1 Generic gender classifier
We collect 2400 Mongoloid males and 2500 Mongoloid females, 2400 Caucasoid
males and 2400 Caucasoid females, and 1800 African males and 1600 African fe-
males from consumer images for training. Another independent test set is collected
which contains 400 faces for each ethnicity with half for each gender. We train two
kinds of gender classifiers: first we train gender classifier for each ethnicity respec-
tively and the results on test set are show in Table 3; and second, we train a gender
classifier using all the males and females in the training set and the results on test set
are shown in Table 4. All the gender classifiers in this section are trained in the same
way as in Section 3.
Table.3. Gender classifier for each ethnicity respectively (MC, CC and AC mean gender
classifier on Mongoloid, Caucasoid and African respectively)
Mongoloid Caucasoid African
Male Female Male Female Male Female
MC 90.1% 92.9% 95.7% 61.3% 93.5% 50%
CC 78.5% 89.1% 96.8% 88.7% 94% 69%
AC 52% 95.5% 55.5% 96% 94% 82%
Table.4. Generic gender classifier for all ethnicities
Mongoloid Caucasoid African
Male Female Male Female Male Female
84.5% 93.5% 86% 93% 95.5% 77%
-0.1 -0.08 -0.06 -0.04 -0.02 00.02 0.04 0.06 0.08 0.1
0
5
10
15
20
25
30
35
40
45
50
confidence
sample distribution
Gender distribution in muti-races environment
Mon male
Mon female
Cau male
Cau female
Afr male
Afr female
Fig.4. Confidence distribution for different ethnicity in generic gender classifier (Positive
for male and negative for female)
We can conclude from Table 3 that the gender classifier behaves well on the eth-
nicity it is trained on while can't achieve good results on other ethnicities. When we
train a generic gender classifier for all ethnicities, the result as in Table 4 is not as
good as training specific gender classifier as in Table 3. This can be explained by
Fig.4, in generic gender classifier, we try to find the same threshold for all ethnicity
faces, which in fact is not a best decision boundary for each ethnicity. As show in
Fig.4, the decision boundary for Africans is apt to male side while the decision
boundary for Mongoloid is apt to female side. That is why the generic gender classi-
fier is inclined to classify Africans as males and Mongoloid as females as shown in
Table 4.
4.2 Ethnicity specific gender classifier
Enlightened by the analysis in Section 4.1, we propose an ethnicity specific gender
classification framework as shown in Fig.5 for multiethnic environment. In the new
framework, the ethnicity is treated as a latent variable for gender classification. We
can formalize the procedure as:
(|) (|,)(|)
E
PG F PG EFPE F=
where G, E and F represent gender, ethnicity and face respectively.
We trained an ethnicity classifier with samples collected in Section 4.1 using
AdaBoost.MH [20] and Haar-like features [16]. Gender classifiers on each ethnicity
are from Section 4.1. The results of ethnicity specific gender classifier are compared
with the generic gender classifier in Table 5. We can see that the ethnicity specific
gender classifier performs better than the generic gender classifier, especially on
Mongoloid males and African females, which is consistent with analysis of Fig.4.
This experiment hints that faces from different ethnicity have different gender feature
and in a multiethnic environment, gender classifier could be better by taking ethnicity
as a latent variable. Some results are shown in Fig.6.
Fig.5. Ethnicity specific gender classification framework
Table.5 Comparison of Generic Gender Classifier (GGC) and Ethnicity Specific Gender
Classifier (ESGC)
Mongoloid Caucasoid African
male female male female male female
GGC 84.5% 93.5% 86% 93% 95.5% 77%
ESGC 89% 93% 86.3% 96.6% 94% 82%
Fig.6. Gender classification results on consumer images in multiethnic environment
5 Conclusion
In this paper, a PBT approach for face gender classification on consumer images is
presented. Faces on consumer images vary greatly in pose, illumination and expres-
sion that make it much more difficult than in the constrained environments. In this
approach, Active Shape Model (ASM) is used for normalization and a PBT is trained
for classification by which through divide and conquer strategy a more accurate clas-
sification boundary on consumer images is achieved. Experiments on both snapshots
and consumer images show that the PBT method is better than the SVM and Real
AdaBoost methods.
We also discussed the ethnicity factor in gender classification experimentally, to
our best knowledge there is no such work before. We find that faces from different
ethnicity have different gender feature, and gender classifier trained on a specific
ethnicity could not get good generalization ability on other ethnicities. Finally, we
improve the performance of gender classification in a multiethnic environment by
treating ethnicity as a latent variable.
However, currently we can only deal with frontal or near frontal faces from con-
sumer images. And the accuracy of gender classifier on Africans is not as high as on
Mongoloid and Caucasoid. Another issue we have not considered is the impact of age
on the face gender classification. Those are our future work.
6 Acknowledgement
This work is supported by National Science Foundation of China under grant
No.60673107, and it is also supported by a grant from HP Corporation.
References
1. B.A. Golomb, D.T. Lawrence, and T. J. Sejnowski: "SEXNET: A Neural Network Identifies
Sex from Human Faces". NIPS 1990.
2. S. Gutta, H. Wechsler, and P.J. Phillips: Gender and Ethnic Classification of Face Images.
FG 1998.
3. K. Balci, V. Atalay: PCA for Gender Estimation: Which Eigenvectors Contribute? ICPR
2002
4. B. Moghaddam, and M.H. Yang: Learning Gender with Support Faces, IEEE Trans. on
Pattern Analysis and Machine Intelligence, Vol.24, No.5, May 2002
5. J. Hayashi, M. Yasumoto, H. Ito, and H. Koshimizu: Age and Gender Estimation based on
Wrinkle Texture. ICPR 2002
6. C. BenAbdelkader, P. Griffin: A Local Region-based Approach to Gender Classification
from Face Images. CVPR 2005
7. Z. Yang, M. Li, and H. Ai: An Experimental Study on Automatic Face Gender Classification.
ICPR 2006
8. G. Shakhnarovich, P.A. Viola, and B. Moghaddam: A Unified Learning Framework for Real
Time Face Detection and Classification. AFG 2002
9. A. Lapedriza, D. Masip, and J. Vitria`: Are External Face Features Useful for Automatic
Face Classification. CVPR 2005
10. A. Lapedriza, M.J. Manuel J. M.-Jimenez, and Jordi Vitria`: Gender Recognition in Non
Controlled Environments. ICPR 2006.
11. Z. Tu: Probabilistic Boosting-Tree: Learning Discriminative Models for Classification,
Recognition, and Clustering. ICCV 2005.
12. C. Huang, H. Ai, B. Wu, and S. Lao: Boosting nested cascade detector for multi-view face
detection. ICPR 2004.
13. L. Zhang, H. Ai, S. Xin, C. Huang, S. Tsukiji, and S. Lao: Robust Face Alignment Based
on Local Texture Classifiers. ICIP 2005.
14. B. Wu, H. Ai, and C. Huang: LUT-Based AdaBoost for Gender Classification. In AVBPA
2003.
15. R.E. Schapire, Y. Singer: Improved Boosting Algorithms Using Confidence-rated Predic-
tions. Machine Learning 37, 297–336, (1999)
16. P. Viola and M. Jones: Fast Multi-view Face Detection. Proc. of CVPR, 2001.
17. S. Gutta, J. R. Huang, P. Jonathon, and H. Wechsler: Mixture of Experts for Classification
of Gender, Ethnic Origin, and Pose of Human Faces. IEEE Transactions on Neural Net-
works.
18. Z. Yang, and H. Ai: Demographic Classification with Local Binary Patterns. ICB 2007
19. S. Hosoi, E. Takikawa, and M. Kawade: Ethnicity Estimation with Facial Images. FG 2004
20. R. E. Schapire, and Y. Singer: Improved Boosting Algorithms Using Confidence-rated
Predictions. Machine Learning 1999.
... This would lead to misrepresentation rather than generalisation. A typical example of this type of bias is in regards to gender [231]. In [232] the authors found, for example, that models tend to have a bias towards a particular gender even when a dataset is balanced. ...
Thesis
This thesis is focused on the application of computer audition (i. e., machine listening) methodologies for monitoring states of emotional wellbeing. Computer audition is a growing field and has been successfully applied to an array of use cases in recent years. There are several advantages to audio-based computational analysis; for example, audio can be recorded non-invasively, stored economically, and can capture rich information on happenings in a given environment, e. g., human behaviour. With this in mind, maintaining emotional wellbeing is a challenge for humans and emotion-altering conditions, including stress and anxiety, have become increasingly common in recent years. Such conditions manifest in the body, inherently changing how we express ourselves. Research shows these alterations are perceivable within vocalisation, suggesting that speech-based audio monitoring may be valuable for developing artificially intelligent systems that target improved wellbeing. Furthermore, computer audition applies machine learning and other computational techniques to audio understanding, and so by combining computer audition with applications in the domain of computational paralinguistics and emotional wellbeing, this research concerns the broader field of empathy for Artificial Intelligence (AI). To this end, speech-based audio modelling that incorporates and understands paralinguistic wellbeing-related states may be a vital cornerstone for improving the degree of empathy that an artificial intelligence has. To summarise, this thesis investigates the extent to which speech-based computer audition methodologies can be utilised to understand human emotional wellbeing. A fundamental background on the fields in question as they pertain to emotional wellbeing is first presented, followed by an outline of the applied audio-based methodologies. Next, detail is provided for several machine learning experiments focused on emotional wellbeing applications, including analysis and recognition of under-researched phenomena in speech, e. g., anxiety, and markers of stress. Core contributions from this thesis include the collection of several related datasets, hybrid fusion strategies for an emotional gold standard, novel machine learning strategies for data interpretation, and an in-depth acoustic-based computational evaluation of several human states. All of these contributions focus on ascertaining the advantage of audio in the context of modelling emotional wellbeing. Given the sensitive nature of human wellbeing, the ethical implications involved with developing and applying such systems are discussed throughout.
... Edge and texture feature provide the sufficient information for unique determination of the gender. histogram of gradient (HOG) [5], local binary pattern [6] and CNN provide the texture pattern of face while Gabor [4], DCT [7], scale invariant feature [8] etc provide the edge feature of face. Dimension of feature are reduced using dimensionality reduction and features are optimized using feature optimization approach. ...
Article
Full-text available
Facial Gender Analysis has application of specific gender entry detection, human machine interface for digital marketing, real time targeted advertisement and gender demo-graphic analysis. The facial gender can be predicted by classifi-cation of the texture and unique edges pattern. Gabor filter can extract the edge- texture patterns on the face but has problem of high dimensionality with redundancy. For accuracy enhance-ment, the dimension and redundancy is needed to reduce by proposed technique as maxDWT feature optimization method. The proposed model is evaluated on real life challenging dataset of face as illumination variation, POSE, face profile, age variation and obstruction on face as hat, birthmark, moles, speckles, beard, etc. Results shows that proposed technique far better than existing state of art methods of gender prediction.
... Face analysis is one of the most widely used subdisciplines of machine vision as a lot of information can be extracted from the face. For example, by having a person's face, a model can determine that person's gender [1][2][3][4][5], estimate his/her age [6][7][8][9], understand the person's emotional circumstance [10][11][12][13], and even realize the existence of some diseases [14][15]. ...
Preprint
Full-text available
The main idea of this paper is that if a model can recognize a person, of course, it must be able to know the gender of that person, too. Therefore, instead of defining a new model for gender classification, this paper uses ArcFace features to determine gender, based on the facial features. A face image is given to ArcFace and 512 features are obtained for the face. Then, with the help of traditional machine learning models, gender is determined. Discriminative methods such as Support Vector Machine (SVM), Linear Discriminant, and Logistic Regression well demonstrate that the features extracted from the ArcFace create a remarkable distinction between the gender classes. Experiments on the Gender Classification Dataset show that SVM with Gaussian kernel is able to classify gender with an accuracy of 96.4% using ArcFace features.
... Results in [40] show that the classification rate tends to decrease in an unconstrained environment when compared to a constrained environment. The majority of previous research works on gender estimation have used to face datasets collected in constrained environments [41]. ...
Conference Paper
Full-text available
Recently, biometric research has advanced substantially, especially in the recognition of unfamiliar faces through automated face recognition technology. Face recognition technology has gained popularity in a lot of applications, top of which is security and law enforcement. As an emerging trend that has not gained the expected level of acceptance due to its potential to violate privacy, a lot of research has been ongoing to better improve the technology. While a lot of countries and regions have made great efforts to improving and legislating the use of this technology, a number of developing countries are still lagging. The objective of this study is to provide a survey on face technology for researchers in developing countries like Nigeria to be updated with research concerns and gap-filling options in a technology that will become the baseline for global security and immigration in the nearest future. A survey on face recognition methods, face presentation attack detection, and facial soft biometric traits are presented in this paper.
Chapter
Facial attributes indicate the intuitive semantic descriptions of a human face like gender, race, expression, and so on. In the past few years, automated facial attribute analysis has become an active field in the area of biometric recognition due to its wide range of possible applications, such as face verification [5, 59], face identification [63, 80], or surveillance [110], just to mention a few.
Article
Full-text available
This paper extends the face detection framework proposed by Viola and Jones 2001 to handle profile views and rotated faces. As in the work of Rowley et al. 1998, and Schneiderman et al. 2000, we build different detectors for different views of the face. A decision tree is then trained to determine the viewpoint class (such as right profile or rotated 60 degrees) for a given window of the image being examined. This is similar to the approach of Rowley et al. 1998. The appropriate detector for that viewpoint can then be run instead of running all detectors on all windows. This technique yields good results and maintains the speed advantage of the Viola-Jones detector. This work may not be copied or reproduced in whole or in part for any commercial purpose. Permission to copy in whole or in part without payment of fee is granted for nonprofit educational and research purposes provided that all such whole or partial copies include the following: a notice that such copying is by permission of Mitsubishi Electric Research Laboratories, Inc.; an acknowledgment of the authors and individual contributions to the work; and all applicable portions of the copyright notice. Copying, reproduction, or republishing for any other purpose shall require a license with payment of fee to Mitsubishi Electric Research Laboratories, Inc. All rights reserved.
Conference Paper
Full-text available
LBP (Local Binary Pattern) as an image operator is used to extract LBPH (LBP histogram) features for texture description. In this paper, we present a novel method to use LBPH feature in ordinary binary classiflcation problem. Given a restricted local patch, the Chi square distance between the extracted LBPH and a reference histogram is used as a measure of confldence belonging to the reference class, and an optimal reference histogram is obtained by iteratively optimization; real AdaBoost algorithm is used to learn a sequence of best local features iteratively and combine them into a strong classifler. The experiments on age, gender and ethnicity classiflcation demonstrate its efiectiveness.
Conference Paper
Full-text available
This paper presents an experimental study on auto- matic face gender classification by building a system that mainly consists of four parts, face detection, face alignment, texture normalization and gender classifi- cation. Comparative study on the effects of different texture normalization methods including two kinds of affine mapping and one Delaunay triangulation based warping as preprocesses for gender classification by SVM, LDA and Real Adaboost respectively is reported through experiments on very large sets of snapshot images.
Conference Paper
This paper presents progress toward an integrated, robust, real-time face detection and demographic analysis system. Faces are detected and extracted using the fast algorithm proposed by P. Viola and M.J. Jones (2001). Detected faces are passed to a demographic (gender and ethnicity) classifier which uses the same architecture as the face detector. This demographic classifier is extremely fast, and delivers error rates slightly better than the best-known classifiers. To counter the unconstrained and noisy sensing environment, demographic information is integrated across time for each individual. Therefore, the final demographic classification combines estimates from many facial detections in order to reduce the error rate. The entire system processes 10 frames per second on an 800-MHz Intel Pentium III
Article
In this paper a new experiment using the FRGC database is proposed. The experiment deals with the use of external face features for face classification. Unlike the most part of algorithms that can be found in the literature for classifying faces, we consider the external information located at hair and ears as a reliable source of information. These features have often been discarded due to the difficulty of their extraction and alignment, and the lack of robustness in security related applications. Nevertheless, there are a lot of applications where these considerations are not valid, and the proper processing of external features can be an important additional source of information for classifications tasks. We also propose, following this assumption, a method for extracting external information from face images. The method is based on a top-down reconstructionbased algorithm for extracting the external face features. Once extracted, they are encoded in a second step using the Non Negative Matrix Factorization (NMF) algorithm, yielding an aligned high dimensional feature vector. This method has been used in a gender recognition problem, concluding that the encoded information is useful for classification purposes.
Conference Paper
There are two main approaches to the problem of gender classification, Support Vector Machines (SVMs) and Adaboost learning methods, of which SVMs are better in correct rate but are more computation intensive while Adaboost ones are much faster with slightly worse performance. For possible real-time applications the Adaboost method seems a better choice. However, the existing Adaboost algorithms take simple threshold weak classifiers, which are too weak to fit complex distributions, as the hypothesis space. Because of this limitation of the hypothesis model, the training procedure is hard to converge. This paper presents a novel Look Up Table (LUT) weak classifier based Adaboost approach to learn gender classifier. This algorithm converges quickly and results in efficient classifiers. The experiments and analysis show that the LUT weak classifiers are more suitable for boosting procedure than threshold ones.
Conference Paper
In most of the automatic face classication applications, images should be captured in natural environments, where partial occlusions or high local changes in the illumination are frequent. For this reason, face classication tasks in uncontrolled environment are still nowadays unsolved prob- lems, given that the loss of information caused by these ar- tifacts can easily mislead any classier . We present in this paper a system to extract robust face features that can be applied to encode information from any zone of the face and that can be used for different face classication prob- lems. To test this method we include the results obtained in different gender classication experiments, considering controlled and uncontrolled environments and extracting face features from internal and external face zones. The obtained rates show, on the one hand, that we can obtain signicant information applying the presented feature ex- traction scheme and, on the other hand, that the external face zone can contribute useful information for classica- tion purposes.