Dr Wenhao Zhang Wenhao.Zhang@uwe.ac.uk
Associate Professor of Computer Vision and Machine Learning
Gender recognition from facial images: Two or three dimensions?
Zhang, Wenhao; Smith, Melvyn L.; Smith, Lyndon N.; Farooq, Abdul
Authors
Melvyn Smith Melvyn.Smith@uwe.ac.uk
Research Centre Director Vision Lab/Prof
Lyndon Smith Lyndon.Smith@uwe.ac.uk
Professor in Computer Simulation and Machine
Abdul Farooq Abdul2.Farooq@uwe.ac.uk
Associate Director (Human-Centric Robotics)
Abstract
© 2016 Optical Society of America. This paper seeks to compare encoded features from both two-dimensional (2D) and three-dimensional (3D) face images in order to achieve automatic gender recognition with high accuracy and robustness. The Fisher vector encoding method is employed to produce 2D, 3D, and fused features with escalated discriminative power. For 3D face analysis, a two-source photometric stereo (PS) method is introduced that enables 3D surface reconstructions with accurate details as well as desirable efficiency. Moreover, a 2D + 3D imaging device, taking the two-source PS method as its core, has been developed that can simultaneously gather color images for 2D evaluations and PS images for 3D analysis. This system inherits the superior reconstruction accuracy from the standard (three or more light) PS method but simplifies the reconstruction algorithm as well as the hardware design by only requiring two light sources. It also offers great potential for facilitating human computer interaction by being accurate, cheap, efficient, and nonintrusive. Ten types of low-level 2D and 3D features have been experimented with and encoded for Fisher vector gender recognition. Evaluations of the Fisher vector encoding method have been performed on the FERET database, Color FERET database, LFW database, and FRGCv2 database, yielding 97.7%, 98.0%, 92.5%, and 96.7% accuracy, respectively. In addition, the comparison of 2D and 3D features has been drawn from a self-collected dataset, which is constructed with the aid of the 2D + 3D imaging device in a series of data capture experiments. With a variety of experiments and evaluations, it can be proved that the Fisher vector encoding method outperforms most state-of-the-art gender recognition methods. It has also been observed that 3D features reconstructed by the two-source PS method are able to further boost the Fisher vector gender recognition performance, i.e., up to a 6% increase on the self-collected database.
Journal Article Type | Article |
---|---|
Acceptance Date | Dec 18, 2015 |
Online Publication Date | Mar 1, 2016 |
Publication Date | Mar 1, 2016 |
Deposit Date | Feb 8, 2016 |
Publicly Available Date | Apr 6, 2016 |
Journal | Journal of the Optical Society of America A: Optics and Image Science, and Vision |
Print ISSN | 1084-7529 |
Electronic ISSN | 1520-8532 |
Publisher | Optical Society of America |
Peer Reviewed | Peer Reviewed |
Volume | 33 |
Issue | 3 |
Pages | 333-344 |
DOI | https://doi.org/10.1364/JOSAA.33.000333 |
Keywords | gender recognition, computers, 2D, 3D |
Public URL | https://uwe-repository.worktribe.com/output/909684 |
Publisher URL | http://dx.doi.org/10.1364/JOSAA.33.000333 |
Contract Date | Feb 11, 2016 |
Files
Gender recognition from facial images 2D or 3D.pdf
(1.4 Mb)
PDF
You might also like
Eye center localization and gaze gesture recognition for human-computer interaction
(2016)
Journal Article
Predicting multiple-target search performance using eye movements and individual differences
(2023)
Presentation / Conference Contribution
Eye centre localisation: An unsupervised modular approach
(2016)
Journal Article
Downloadable Citations
About UWE Bristol Research Repository
Administrator e-mail: repository@uwe.ac.uk
This application uses the following open-source libraries:
SheetJS Community Edition
Apache License Version 2.0 (http://www.apache.org/licenses/)
PDF.js
Apache License Version 2.0 (http://www.apache.org/licenses/)
Font Awesome
SIL OFL 1.1 (http://scripts.sil.org/OFL)
MIT License (http://opensource.org/licenses/mit-license.html)
CC BY 3.0 ( http://creativecommons.org/licenses/by/3.0/)
Powered by Worktribe © 2024
Advanced Search