Abstract

In many real-world face recognition scenarios, face images can hardly be aligned accurately due to complex appearance variations or low-quality images. To address this issue, we propose a new approach to extract robust face region descriptors. Specifically, we divide each image (resp. video) into several spatial blocks (resp. spatial-temporal volumes) and then represent each block (resp. volume) by sum-pooling the nonnegative sparse codes of position-free patches sampled within the block (resp. volume). Whitened Principal Component Analysis (WPCA) is further utilized to reduce the feature dimension, which leads to our Spatial Face Region Descriptor (SFRD) (resp. Spatial-Temporal Face Region Descriptor, STFRD) for images (resp. videos). Moreover, we develop a new distance metric learning method for face verification called Pairwise-constrained Multiple Metric Learning (PMML) to effectively integrate the face region descriptors of all blocks (resp. volumes) from an image (resp. a video). Our work achieves the state-of-the-art performances on two real-world datasets LFW and YouTube Faces (YTF) according to the restricted protocol.

Keywords

Artificial intelligencePattern recognition (psychology)Facial recognition systemComputer scienceMetric (unit)Face (sociological concept)Pairwise comparisonPoolingBlock (permutation group theory)Computer visionPrincipal component analysisFeature (linguistics)Feature extractionMathematics

Affiliated Institutions

Related Publications

Publication Info

Year
2013
Type
article
Pages
3554-3561
Citations
194
Access
Closed

External Links

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

194
OpenAlex

Cite This

Zhen Cui, Li Wen, Dong Xu et al. (2013). Fusing Robust Face Region Descriptors via Multiple Metric Learning for Face Recognition in the Wild. , 3554-3561. https://doi.org/10.1109/cvpr.2013.456

Identifiers

DOI
10.1109/cvpr.2013.456