75.3%
5.1. technique presentation.
The hybrid approaches are based on local and subspace features in order to use the benefits of both subspace and local techniques, which have the potential to offer better performance for face recognition systems.
Flowchart of the proposed multimodal deep face representation (MM-DFR) technique [ 95 ]. CNN, convolutional neural network.
The proposed CNN–LSTM–ELM [ 103 ].
Table 3 summarizes the hybrid approaches that we presented in this section. Various techniques are introduced to improve the performance and the accuracy of recognition systems. The combination between the local approaches and the subspace approach provides robust recognition and reduction of dimensionality under different illumination conditions and facial expressions. Furthermore, these technologies are presented to be sensitive to noise, and invariant to translations and rotations.
Hybrid approaches. GW, Gabor wavelet; OCLBP, over-complete LBP; WCCN, within class covariance normalization; WLBP, Walsh LPB; ICP, iterative closest point; LGBPHS, local Gabor binary pattern histogram sequence; FLD, Fisher linear discriminant; SAE, stacked auto-encoder.
Author/Technique Used | Database | Matching | Limitation | Advantage | Result | |
---|---|---|---|---|---|---|
Fathima et al. [ ] | GW-LDA | AT&T | k-NN | High processing time | Illumination invariant and reduce the dimensionality | 88% |
FACES94 | 94.02% | |||||
MITINDIA | 88.12% | |||||
Barkan et al., [ ] | OCLBP, LDA, and WCCN | LFW | WCCN | _ | Reduce the dimensionality | 87.85% |
Juefei et al. [ ] | ACF and WLBP | LFW | Complexity | Pose conditions | 89.69% | |
Simonyan et al. [ ] | Fisher + SIFT | LFW | Mahalanobis matrix | Single feature type | Robust | 87.47% |
Sharma et al. [ ] | PCA–ANFIS | ORL | ANFIS | Sensitivity-specificity | 96.66% | |
ICA–ANFIS | ANFIS | Pose conditions | 71.30% | |||
LDA–ANFIS | ANFIS | 68% | ||||
Ojala et al. [ ] | DCT–PCA | ORL | Euclidian distance | Complexity | Reduce the dimensionality | 92.62% |
UMIST | 99.40% | |||||
YALE | 95.50% | |||||
Mian et al. [ ] | Hotelling transform, SIFT, and ICP | FRGC | ICP | Processing time | Facial expressions | 99.74% |
Cho et al. [ ] | PCA–LGBPHS | Extended Yale Face | Bhattacharyya distance | Illumination condition | Complexity | 95% |
PCA–GABOR Wavelets | ||||||
Sing et al. [ ] | PCA–FLD | CMU | SVM | Robustness | Pose, illumination, and expression | 71.98% |
FERET | 94.73% | |||||
AR | 68.65% | |||||
Kamencay et al. [ ] | SPCA-KNN | ESSEX | KNN | Processing time | Expression variation | 96.80% |
Sun et al. [ ] | CNN–LSTM–ELM | OPPORTUNITY | LSTM/ELM | High processing time | Automatically learn feature representations | 90.60% |
Ding et al. [ ] | CNNs and SAE | LFW | _ _ | Complexity | High recognition rate | 99% |
In the last step of recognition, the face extracted from the background during the face detection step is compared with known faces stored in a specific database. To make the decision, several techniques of comparison are used. This section describes the most common techniques used to make the decision and comparison.
In general, the Euclidean distance between two points P = ( 1 , p 2 , … , p n ) and Q = ( q 1 , q 2 , … , q n ) in the n-dimensional space would be defined by the following:
There are many face classification techniques in the literature that allow to select, from a few examples, the group or class to which the objects belong. Some of them are based on statistics, such as the Bayesian classifier and correlation [ 18 ], and so on, and others based on the regions that generate the different classes in the decision space, such as K-means [ 9 ], CNN [ 103 ], artificial neural networks (ANNs) [ 37 ], support vector machines (SVMs) [ 26 , 107 ], k-nearest neighbors (K-NNs), decision trees (DTs), and so on.
Optimal hyperplane, support vectors, and maximum margin.
There is an infinite number of hyperplanes capable of perfectly separating two classes, which implies to select a hyperplane that maximizes the minimal distance between the learning examples and the learning hyperplane (i.e., the distance between the support vectors and the hyperplane). This distance is called “margin”. The SVM classifier is used to calculate the optimal hyperplane that categorizes a set of labels training data in the correct class. The optimal hyperplane is solved as follows:
Given that x i are the training features vectors and y i are the corresponding set of l (1 or −1) labels. An SVM tries to find a hyperplane to distinguish the samples with the smallest errors. The classification function is obtained by calculating the distance between the input vector and the hyperplane.
where w and b are the parameters of the model. Shen et al. [ 108 ] proposed the Gabor filter to extract the face features and applied the SVM for classification. The proposed FaceNet method achieves a good record accuracy of 99.63% and 95.12% using the LFW YouTube Faces DB datasets, respectively.
Artificial neural network.
Various variants of neural networks have been developed in the last years, such as convolutional neural networks (CNN) [ 14 , 110 ] and recurrent neural networks (RNN) [ 111 ], which very effective for image detection and recognition tasks. CNNs are a very successful deep model and are used today in many applications [ 112 ]. From a structural point of view, CNNs are made up of three different types of layers: convolution layers, pooling layers, and fully-connected layers.
Wen et al. [ 113 ] introduce a new supervision signal, called center loss, for the face recognition task in order to improve the discriminative power of the deeply learned features. Specifically, the proposed center loss function is trainable and easy to optimize in the CNNs. Several important face recognition benchmarks are used for evaluation including LFW, YTF, and MegaFace Challenge. Passalis and Tefas [ 114 ] propose a supervised codebook learning method for the bag-of-features representation able to learn face retrieval-oriented codebooks. This allows using significantly smaller codebooks enhancing both the retrieval time and storage requirements. Liu et al. [ 115 ] and Amato et al. [ 116 ] propose a deep face recognition technique under open-set protocol based on the CNN technique. A face dataset composed of 39,037 faces images belonging to 42 different identities is used to perform the experiments. Taigman et al. [ 117 ] present a system (DeepFace) able to outperform existing systems with only very minimal adaptation. It is trained on a large dataset of faces acquired from a population vastly different than the one used to construct the evaluation benchmarks. This technique achieves an accuracy of 97.35% on the LFW. Ma et al. [ 118 ] introduce a robust local binary pattern (LBP) guiding pooling (G-RLBP) mechanism to improve the recognition rates of the CNN models, which can successfully lower the noise impact. Koo et al. [ 119 ] propose a multimodal human recognition method that uses both the face and body and is based on a deep CNN. Cho et al. [ 120 ] propose a nighttime face detection method based on CNN technique for visible-light images. Koshy and Mahmood [ 121 ] develop deep architectures for face liveness detection that uses a combination of texture analysis and a CNN technique to classify the captured image as real or fake. Elmahmudi and Ugail [ 122 ] present the performance of machine learning for face recognition using partial faces and other manipulations of the face such as rotation and zooming, which we use as training and recognition cues. The experimental results on the tasks of face verification and face identification show that the model obtained by the proposed DNN training framework achieves 97.3% accuracy on the LFW database with low training complexity. Seibold et al. [ 123 ] proposed a morphing attack detection method based on DNNs. A fully automatic face image morphing pipeline with exchangeable components was used to generate morphing attacks, train neural networks based on these data, and analyze their accuracy. Yim et al. [ 124 ] propose a new deep architecture based on a novel type of multitask learning, which can achieve superior performance in rotating to a target-pose face image from an arbitrary pose and illumination image while preserving identity. Nguyen et al. [ 111 ] propose a new approach for detecting presentation attack face images to enhance the security level of a face recognition system. The objective of this study was the use of a very deep stacked CNN–RNN network to learn the discrimination features from a sequence of face images. Finally, Bajrami et al. [ 125 ] present experiment results with LDA and DNN for face recognition, while their efficiency and performance are tested on the LFW dataset. The experimental results show that the DNN method achieves better recognition accuracy, and the recognition time is much faster than that of the LDA method in large-scale datasets.
The most commonly used databases for face recognition systems under different conditions are Pointing Head Pose Image Database (PHPID) [ 126 ], Labeled Faces in Wild (LFW) [ 127 ], FERET [ 15 , 16 ], ORL, and Yale. The last are used for face recognition systems under different conditions, which provide information for supervised and unsupervised learning. Supervised learning is based on two training modules: image unrestricted training setting and image restricted training setting. For the first model, only “same” or “not same” binary labels are used in the training splits. For the second model, the identities of the person in each pair are provided in the training splits.
In this section, we present some advantages and disadvantages of holistic, local, and hybrid approaches to identifying faces during the last 20 years. DL approaches can be considered as a statistical approach (holistic method), because the training procedure scheme usually searches for statistical structures in the input patterns. Table 4 presents a brief summary of the three approaches.
General performance of face recognition approaches.
Approaches | Databases Used | Advantages | Disadvantages | Performances | Challenges Handled | |
---|---|---|---|---|---|---|
TDF, CF1999, LFW, FERET, CMU-PIE, AR, Yale B, PHPID, YaleB Extended, FRGC2.0, Face94. | ]. , ]. | , ]. | ], various lighting conditions[ ], facial expressions [ ], and low resolution. | |||
]. ]. | ]. | ]. ]. | ]. | |||
LFW, FERET, MEPCO, AR, ORL, CK, MMI, JAFFE, C. Yale B, Yale, MNIST, ORL, UMIST face, HELEN face, FRGC. | , ]. , , , ]. | ]. | ]. ]. | , ], scaling, facial expressions. | ||
, , ]. , , ]. | ]. , ]. | ]. , ]. | , ], poses [ ], conditions, scaling, facial expressions. | |||
AT&T, FACES94, MITINDIA, LFW, ORL, UMIST, YALE, FRGC, Extended Yale, CMU, FERET, AR, ESSEX. | ]. | , , ]. | ]. ]. | , ]. |
7.1. discussion.
In the past decade, the face recognition system has become one of the most important biometric authentication methods. Many techniques are used to develop many face recognition systems based on facial information. Generally, the existing techniques can be classified into three approaches, depending on the type of desired features.
In particular, recognition methods performed on static images produce good results under different lighting and expression conditions. However, in most cases, only the face images are processed at the same size and scale. Many methods require numerous training images, which limits their use for real-time systems, where the response time is an important aspect.
The main purpose of techniques such as HOG, LBP, Gabor filters, BRIEF, SURF, and SIFT is to discover distinctive features, which can be divided into two parts: (1) local appearance-based techniques, which are used to extract local features when the face image is divided into small regions (including HOG, LBP, Gabor filters, and correlation filters); and (2) key-points-based techniques, which are used to detect the points of interest in the face image, after which features’ extraction is localized based on these points, including BRIEF, SURF, and SIFT. In the context of face recognition, local techniques only treat certain facial features, which make them very sensitive to facial expressions and occlusions [ 4 , 14 , 37 , 50 , 51 , 52 , 53 ]. The relative robustness is the main advantage of these feature-based local techniques. Additionally, they take into account the peculiarity of the face as a natural form to recognize a reduced number of parameters. Another advantage is that they have a high compaction capacity and a high comparison speed. The main disadvantages of these methods are the difficulty of automating the detection of facial features and the fact that the person responsible for the implementation of these systems must make an arbitrary decision on really important points.
Unlike the local approaches, holistic approaches are other methods used for face recognition, which treat the whole face image and do not require extracting face regions or features points (eyes, mouth, noses, and so on). The main function of these approaches is to represent the face image with a matrix of pixels. This matrix is often converted into feature vectors to facilitate their treatment. After that, the feature vectors are applied in a low-dimensional space. In fact, subspace techniques are sensitive to different variations (facial expressions, illumination, and different poses), which make them easy to implement. Many subspace techniques are implemented to represent faces such as Eigenface, Eigenfisher, PCA, and LDA, which can be divided into two categories: linear and non-linear techniques. The main advantage of holistic approaches is that they do not destroy image information by focusing only on regions or points of interest. However, this property represents a disadvantage because it assumes that all the pixels of the image have the same importance. As a result, these techniques are not only computationally expensive, but also require a high degree of correlation between the test and the training images. In addition, these approaches generally ignore local details, which means they are rarely used to identify faces.
Hybrid approaches are based on local and global features to exploit the benefits of both techniques. These approaches combine the two approaches described above into a single system to improve the performance and accuracy of recognition. The choice of the required method to be used must take into account the application in which it was applied. For example, in the face recognition systems that use very small images, methods based on local features are a bad choice. Another consideration in the algorithm selection process is the number of training examples needed. Finally, we can remember that the tendency is to develop hybrid methods that combine the advantages of local and holistic approaches, but these methods are very complex and require more processing time.
A notable limitation that we found in all the publications reviewed is methodological: despite that the 2D facial recognition has reached a significant level of maturity and a high success rate, it is not surprising that it continues to be one of the most active research areas in computer vision. Considering the results published to date, in the opinion of these authors, three particularly promising techniques for further development of this area stand out: (i) the development of 3D face recognition methods; (ii) the use of multimodal fusion methods of complementary data types, in particular those based on visible and infrared images; and (iii) the use of DL methods.
Finally, researchers have gone further by using multimodal and DL facial recognition systems.
Face recognition system is a popular study task in the field of image processing and computer vision, owing to its potentially enormous application as well as its theoretical value. This system is widely deployed in many real-world applications such as security, surveillance, homeland security, access control, image search, human-machine, and entertainment. However, these applications pose different challenges such as lighting conditions and facial expressions. This paper highlights the recent research on the 2D or 3D face recognition system, focusing mainly on approaches based on local, holistic (subspace), and hybrid features. A comparative study between these approaches in terms of processing time, complexity, discrimination, and robustness was carried out. We can conclude that local feature techniques are the best choice concerning discrimination, rotation, translation, complexity, and accuracy. We hope that this survey paper will further encourage researchers in this field to participate and pay more attention to the use of local techniques for face recognition systems.
Y.K. highlights the recent research on the 2D or 3D face recognition system, focusing mainly on approaches based on local, holistic, and hybrid features. M.J., A.A.F. and M.A. supervised the research and helped in the revision processes. All authors have read and agreed to the published version of the manuscript.
The paper is co-financed by L@bISEN of ISEN Yncrea Ouest Brest, France, Dept Ai-DE, Team Vision-AD and by FSM University of Monastir, Tunisia with collaboration of the Ministry of Higher Education and Scientific Research of Tunisia. The context of the paper is the PhD project of Yassin Kortli.
The authors declare no conflict of interest.
You are accessing a machine-readable page. In order to be human-readable, please install an RSS reader.
All articles published by MDPI are made immediately available worldwide under an open access license. No special permission is required to reuse all or part of the article published by MDPI, including figures and tables. For articles published under an open access Creative Common CC BY license, any part of the article may be reused without permission provided that the original article is clearly cited. For more information, please refer to https://www.mdpi.com/openaccess .
Feature papers represent the most advanced research with significant potential for high impact in the field. A Feature Paper should be a substantial original Article that involves several techniques or approaches, provides an outlook for future research directions and describes possible research applications.
Feature papers are submitted upon individual invitation or recommendation by the scientific editors and must receive positive feedback from the reviewers.
Editor’s Choice articles are based on recommendations by the scientific editors of MDPI journals from around the world. Editors select a small number of articles recently published in the journal that they believe will be particularly interesting to readers, or important in the respective research area. The aim is to provide a snapshot of some of the most exciting work published in the various research areas of the journal.
Original Submission Date Received: .
Find support for a specific problem in the support section of our website.
Please let us know what you think of our products and services.
Visit our dedicated information section to learn more about MDPI.
Face recognition systems: a survey.
2.1. essential steps of face recognition systems.
3. local approaches, 3.1. local appearance-based techniques.
4. holistic approach, 4.1. linear techniques.
DCT Algorithm |
where , and |
Kernel PCA Algorithm |
using kernel function: . and normalize with the function: . using kernel function: |
5. hybrid approach, 5.1. technique presentation.
6. assessment of face recognition approaches, 6.1. measures of similarity or distances.
7. discussion about future directions and conclusions, 7.1. discussion.
Author contributions, conflicts of interest.
Click here to enlarge figure
Author/Technique Used | Database | Matching | Limitation | Advantage | Result | |
---|---|---|---|---|---|---|
Local Appearance-Based Techniques | ||||||
Khoi et al. [ ] | LBP | TDF | MAP | Skewness in face image | Robust feature in fontal face | 5% |
CF1999 | 13.03% | |||||
LFW | 90.95% | |||||
Xi et al. [ ] | LBPNet | FERET | Cosine similarity | Complexities of CNN | High recognition accuracy | 97.80% |
LFW | 94.04% | |||||
Khoi et al. [ ] | PLBP | TDF | MAP | Skewness in face image | Robust feature in fontal face | 5.50% |
CF | 9.70% | |||||
LFW | 91.97% | |||||
Laure et al. [ ] | LBP and KNN | LFW | KNN | Illumination conditions | Robust | 85.71% |
CMU-PIE | 99.26% | |||||
Bonnen et al. [ ] | MRF and MLBP | AR (Scream) | Cosine similarity | Landmark extraction fails or is not ideal | Robust to changes in facial expression | 86.10% |
FERET (Wearing sunglasses) | 95% | |||||
Ren et al. [ ] | Relaxed LTP | CMU-PIE | Chisquare distance | Noise level | Superior performance compared with LBP, LTP | 95.75% |
Yale B | 98.71% | |||||
Hussain et al. [ ] | LPQ | FERET/ | Cosine similarity | Lot of discriminative information | Robust to illumination variations | 99.20% |
LFW | 75.30% | |||||
Karaaba et al. [ ] | HOG and MMD | FERET | MMD/MLPD | Low recognition accuracy | Aligning difficulties | 68.59% |
LFW | 23.49% | |||||
Arigbabu et al. [ ] | PHOG and SVM | LFW | SVM | Complexity and time of computation | Head pose variation | 88.50% |
Leonard et al. [ ] | VLC correlator | PHPID | ASPOF | The low number of the reference image used | Robustness to noise | 92% |
Napoléon et al. [ ] | LBP and VLC | YaleB | POF | Illumination | Rotation + Translation | 98.40% |
YaleB Extended | 95.80% | |||||
Heflin et al. [ ] | correlation filter | LFW/PHPID | PSR | Some pre-processing steps | More effort on the eye localization stage | 39.48% |
Zhu et al. [ ] | PCA–FCF | CMU-PIE | Correlation filter | Use only linear method | Occlusion-insensitive | 96.60% |
FRGC2.0 | 91.92% | |||||
Seo et al. [ ] | LARK + PCA | LFW | Cosine similarity | Face detection | Reducing computational complexity | 78.90% |
Ghorbel et al. [ ] | VLC + DoG | FERET | PCE | Low recognition rate | Robustness | 81.51% |
Ghorbel et al. [ ] | uLBP + DoG | FERET | chi-square distance | Robustness | Processing time | 93.39% |
Ouerhani et al. [ ] | VLC | PHPID | PCE | Power | Processing time | 77% |
Lenc et al. [ ] | SIFT | FERET | a posterior probability | Still far to be perfect | Sufficiently robust on lower quality real data | 97.30% |
AR | 95.80% | |||||
LFW | 98.04% | |||||
Du et al. [ ] | SURF | LFW | FLANN distance | Processing time | Robustness and distinctiveness | 95.60% |
Vinay et al. [ ] | SURF + SIFT | LFW | FLANN | Processing time | Robust in unconstrained scenarios | 78.86% |
Face94 | distance | 96.67% | ||||
Calonder et al. [ ] | BRIEF | _ | KNN | Low recognition rate | Low processing time | 48% |
Author/Techniques Used | Databases | Matching | Limitation | Advantage | Result | |
---|---|---|---|---|---|---|
Linear Techniques | ||||||
Seo et al. [ ] | LARK and PCA | LFW | L2 distance | Detection accuracy | Reducing computational complexity | 85.10% |
Annalakshmi et al. [ ] | ICA and LDA | LFW | Bayesian Classifier | Sensitivity | Good accuracy | 88% |
Annalakshmi et al. [ ] | PCA and LDA | LFW | Bayesian Classifier | Sensitivity | Specificity | 59% |
Hussain et al. [ ] | LQP and Gabor | FERET | Cosine similarity | Lot of discriminative information | Robust to illumination variations | 99.2% 75.3% |
LFW | ||||||
Gowda et al. [ ] | LPQ and LDA | MEPCO | SVM | Computation time | Good accuracy | 99.13% |
Z. Cui et al. [ ] | BoW | AR | ASM | Occlusions | Robust | 99.43% |
ORL | 99.50% | |||||
FERET | 82.30% | |||||
Khan et al. [ ] | PSO and DWT | CK | Euclidienne distance | Noise | Robust to illumination | 98.60% |
MMI | 95.50% | |||||
JAFFE | 98.80% | |||||
Huang et al. [ ] | 2D-DWT | FERET | KNN | Pose | Frontal or near-frontal facial images | 90.63% 97.10% |
LFW | ||||||
Perlibakas and Vytautas [ ] | PCA and Gabor filter | FERET | Cosine metric | Precision | Pose | 87.77% |
Hafez et al. [ ] | Gabor filter and LDA | ORL | 2DNCC | Pose | Good recognition performance | 98.33% |
C. YaleB | 99.33% | |||||
Sufyanu et al. [ ] | DCT | ORL | NCC | High memory | Controlled and uncontrolled databases | 93.40% |
Yale | ||||||
Shanbhag et al. [ ] | DWT and BPSO | _ _ | _ _ | Rotation | Significant reduction in the number of features | 88.44% |
Ghorbel et al. [ ] | Eigenfaces and DoG filter | FERET | Chi-square distance | Processing time | Reduce the representation | 84.26% |
Zhang et al. [ ] | PCA and FFT | YALE | SVM | Complexity | Discrimination | 93.42% |
Zhang et al. [ ] | PCA | YALE | SVM | Recognition rate | Reduce the dimensionality | 84.21% |
Fan et al. [ ] | RKPCA | MNIST ORL | RBF kernel | Complexity | Robust to sparse noises | _ |
Vinay et al. [ ] | ORB and KPCA | ORL | FLANN Matching | Processing time | Robust | 87.30% |
Vinay et al. [ ] | SURF and KPCA | ORL | FLANN Matching | Processing time | Reduce the dimensionality | 80.34% |
Vinay et al. [ ] | SIFT and KPCA | ORL | FLANN Matching | Low recognition rate | Complexity | 69.20% |
Lu et al. [ ] | KPCA and GDA | UMIST face | SVM | High error rate | Excellent performance | 48% |
Yang et al. [ ] | PCA and MSR | HELEN face | ESR | Complexity | Utilizes color, gradient, and regional information | 98.00% |
Yang et al. [ ] | LDA and MSR | FRGC | ESR | Low performances | Utilizes color, gradient, and regional information | 90.75% |
Ouanan et al. [ ] | FDDL | AR | CNN | Occlusion | Orientations, expressions | 98.00% |
Vankayalapati and Kyamakya [ ] | CNN | ORL | _ _ | Poses | High recognition rate | 95% |
Devi et al. [ ] | 2FNN | ORL | _ _ | Complexity | Low error rate | 98.5 |
Author/Technique Used | Database | Matching | Limitation | Advantage | Result | |
---|---|---|---|---|---|---|
Fathima et al. [ ] | GW-LDA | AT&T | k-NN | High processing time | Illumination invariant and reduce the dimensionality | 88% |
FACES94 | 94.02% | |||||
MITINDIA | 88.12% | |||||
Barkan et al., [ ] | OCLBP, LDA, and WCCN | LFW | WCCN | _ | Reduce the dimensionality | 87.85% |
Juefei et al. [ ] | ACF and WLBP | LFW | Complexity | Pose conditions | 89.69% | |
Simonyan et al. [ ] | Fisher + SIFT | LFW | Mahalanobis matrix | Single feature type | Robust | 87.47% |
Sharma et al. [ ] | PCA–ANFIS | ORL | ANFIS | Sensitivity-specificity | 96.66% | |
ICA–ANFIS | ANFIS | Pose conditions | 71.30% | |||
LDA–ANFIS | ANFIS | 68% | ||||
Ojala et al. [ ] | DCT–PCA | ORL | Euclidian distance | Complexity | Reduce the dimensionality | 92.62% |
UMIST | 99.40% | |||||
YALE | 95.50% | |||||
Mian et al. [ ] | Hotelling transform, SIFT, and ICP | FRGC | ICP | Processing time | Facial expressions | 99.74% |
Cho et al. [ ] | PCA–LGBPHS | Extended Yale Face | Bhattacharyya distance | Illumination condition | Complexity | 95% |
PCA–GABOR Wavelets | ||||||
Sing et al. [ ] | PCA–FLD | CMU | SVM | Robustness | Pose, illumination, and expression | 71.98% |
FERET | 94.73% | |||||
AR | 68.65% | |||||
Kamencay et al. [ ] | SPCA-KNN | ESSEX | KNN | Processing time | Expression variation | 96.80% |
Sun et al. [ ] | CNN–LSTM–ELM | OPPORTUNITY | LSTM/ELM | High processing time | Automatically learn feature representations | 90.60% |
Ding et al. [ ] | CNNs and SAE | LFW | _ _ | Complexity | High recognition rate | 99% |
Approaches | Databases Used | Advantages | Disadvantages | Performances | Challenges Handled | |
---|---|---|---|---|---|---|
TDF, CF1999, LFW, FERET, CMU-PIE, AR, Yale B, PHPID, YaleB Extended, FRGC2.0, Face94. | ]. , ]. | , ]. | ], various lighting conditions[ ], facial expressions [ ], and low resolution. | |||
]. ]. | ]. | ]. ]. | ]. | |||
LFW, FERET, MEPCO, AR, ORL, CK, MMI, JAFFE, C. Yale B, Yale, MNIST, ORL, UMIST face, HELEN face, FRGC. | , ]. , , , ]. | ]. | ]. ]. | , ], scaling, facial expressions. | ||
, , ]. , , ]. | ]. , ]. | ]. , ]. | , ], poses [ ], conditions, scaling, facial expressions. | |||
AT&T, FACES94, MITINDIA, LFW, ORL, UMIST, YALE, FRGC, Extended Yale, CMU, FERET, AR, ESSEX. | ]. | , , ]. | ]. ]. | , ]. |
Kortli, Y.; Jridi, M.; Al Falou, A.; Atri, M. Face Recognition Systems: A Survey. Sensors 2020 , 20 , 342. https://doi.org/10.3390/s20020342
Kortli Y, Jridi M, Al Falou A, Atri M. Face Recognition Systems: A Survey. Sensors . 2020; 20(2):342. https://doi.org/10.3390/s20020342
Kortli, Yassin, Maher Jridi, Ayman Al Falou, and Mohamed Atri. 2020. "Face Recognition Systems: A Survey" Sensors 20, no. 2: 342. https://doi.org/10.3390/s20020342
Article access statistics, further information, mdpi initiatives, follow mdpi.
Subscribe to receive issue release notifications and newsletters from MDPI journals
1417 Accesses
21 Citations
2 Altmetric
Explore all metrics
Face recognition has long been an active research area in the field of artificial intelligence, particularly since the rise of deep learning in recent years. In some practical situations, each identity has only a single sample available for training. Face recognition under this situation is referred to as single sample face recognition and poses significant challenges to the effective training of deep models. Therefore, in recent years, researchers have attempted to unleash more potential of deep learning and improve the model recognition performance in the single sample situation. While several comprehensive surveys have been conducted on traditional single sample face recognition approaches, emerging deep learning based methods are rarely involved in these reviews. Accordingly, we focus on the deep learning-based methods in this paper, classifying them into virtual sample methods and generic learning methods. In the former category, virtual images or virtual features are generated to benefit the training of the deep model. In the latter one, additional multi-sample generic sets are used. There are three types of generic learning methods: combining traditional methods and deep features, improving the loss function, and improving network structure, all of which are covered in our analysis. Moreover, we review face datasets that have been commonly used for evaluating single sample face recognition models and go on to compare the results of different types of models. Additionally, we discuss problems with existing single sample face recognition methods, including identity information preservation in virtual sample methods, domain adaption in generic learning methods. Furthermore, we regard developing unsupervised methods is a promising future direction, and point out that the semantic gap as an important issue that needs to be further considered.
This is a preview of subscription content, log in via an institution to check access.
Subscribe and save.
Price includes VAT (Russian Federation)
Instant access to the full article PDF.
Rent this article via DeepDyve
Institutional subscriptions
Abdelmaksoud M, Nabil E, Farag I, Hameed HA (2020) A novel neural network method for face recognition with a single sample per person. IEEE Access 8:102212–102221. https://doi.org/10.1109/ACCESS.2020.2999030
Article Google Scholar
Abdolali F, Seyyedsalehi S (2010) Face recognition from a single image per person using deep architecture neural network. In: Proceedings of the 3rd International conference on computer and electrical engineering, vol 1, pp 70–73
Abdolali F, Seyyedsalehi SA (2012) Improving face recognition from a single image per person via virtual images produced by a bidirectional network. Procedia Soc Behav Sci 32:108–116
Adamo A, Grossi G, Lanzarotti R (2012) Sparse representation based classification for face recognition by k-limaps algorithm. In: International conference on image and signal processing. Springer, pp 245–252
Blanz V, Vetter T (2003) Face recognition based on fitting a 3d morphable model. IEEE Trans Pattern Anal Mach Intell 25(9):1063–1074
Bodini M, D’Amelio A, Grossi G, Lanzarotti R, Lin J (2018) Single sample face recognition by sparse recovery of deep-learned lda features. In: International conference on advanced concepts for intelligent vision systems. Springer, pp 297–308
Cai S, Zhang L, Zuo W, Feng X (2016) A probabilistic collaborative representation based approach for pattern classification. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2950–2959
Cao B, Wang N, Li J, Gao X (2019) Data augmentation-based joint learning for heterogeneous face recognition. IEEE Trans Neural Netw Learn Syst 30(6):1731–1743. https://doi.org/10.1109/TNNLS.2018.2872675
Article MathSciNet Google Scholar
Chan TH, Jia K, Gao S, Lu J, Zeng Z, Ma Y (2015) Pcanet: a simple deep learning baseline for image classification? IEEE Trans Image Process 24(12):5017–5032
Article MathSciNet MATH Google Scholar
Chen T, Kornblith S, Norouzi M, Hinton GE (2020) A simple framework for contrastive learning of visual representations. In: Proceedings of the 37th international conference on machine learning, ICML 2020, 13–18 July 2020, Virtual Event, Proceedings of machine learning research. PMLR, vol 119, pp 1597–1607. http://proceedings.mlr.press/v119/chen20j.html
Cheng Y, Zhao J, Wang Z, Xu Y, Jayashree K, Shen S, Feng J (2017) Know you at one glance: a compact vector representation for low-shot learning. In: Proceedings of the IEEE international conference on computer vision workshops, pp 1924–1932
Choe J, Park S, Kim K, Hyun Park J, Kim D, Shim H (2017) Face generation for low-shot learning using generative adversarial networks. In: Proceedings of the IEEE international conference on computer vision workshops, pp 1940–1948
Cuculo V, D’Amelio A, Grossi G, Lanzarotti R, Lin J (2019) Robust single-sample face recognition by sparsity-driven sub-dictionary learning using deep features. Sensors 19(1):146
Deng W, Hu J, Guo J (2012) Extended src: undersampled face recognition via intraclass variant dictionary. IEEE Trans Pattern Anal Mach Intell 34(9):1864–1870
Deng W, Hu J, Guo J (2013) In defense of sparsity based face recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 399–406
Deng W, Hu J, Zhou X, Guo J (2014) Equidistant prototypes embedding for single sample based face recognition with generic learning and incremental learning. Pattern Recogn 47(12):3738–3749
Ding C, Bao T, Karmoshi S, Zhu M (2017) Single sample per person face recognition with kpcanet and a weighted voting scheme. SIViP 11(7):1213–1220
Ding C, Tao D (2017) Trunk-branch ensemble convolutional neural networks for video-based face recognition. IEEE Trans Pattern Anal Mach Intell 40(4):1002–1014
Ding Y, Liu F, Tang Z, Zhang T (2020) Uniform generic representation for single sample face recognition. IEEE Access 8:158281–158292
Ding Z, Guo Y, Zhang L, Fu Y (2018) One-shot face recognition via generative learning. In: 2018 13th IEEE international conference on automatic face & gesture recognition (FG 2018). IEEE, pp 1–7
Ding Z, Guo Y, Zhang L, Fu Y (2019) Generative one-shot face recognition. arXiv:1910.04860
Duan Q, Zhang L (2020) Look more into occlusion: realistic face frontalization and recognition with boostgan. IEEE Trans Neural Netw Learn Syst 32(1):214–228
Ericsson L, Gouk H, Loy CC, Hospedales TM (2021) Self-supervised representation learning: introduction, advances and challenges. CoRR arXiv:abs/2110.09327
Fan L, Sun X, Rosin PL (2021) Siamese graph convolution network for face sketch recognition: An application using graph structure for face photo-sketch recognition. In: 2020 25th international conference on pattern recognition (ICPR). IEEE, pp 8008–8014
Fei-Fei L, Fergus R, Perona P (2006) One-shot learning of object categories. IEEE Trans Pattern Anal Mach Intell 28(4):594–611
Feng Y, Wu F, Shao X, Wang Y, Zhou X (2018) Joint 3d face reconstruction and dense alignment with position map regression network. In: Proceedings of the European conference on computer vision (ECCV), pp 534–551
Galea C, Farrugia RA (2017) Forensic face photo-sketch recognition using a deep learning-based architecture. IEEE Signal Process Lett 24(11):1586–1590
Galea C, Farrugia RA (2017) Matching software-generated sketches to face photographs with a very deep cnn, morphed faces, and transfer learning. IEEE Trans Inf Forensics Secur 13(6):1421–1431
Gao S, Zhang Y, Jia K, Lu J, Zhang Y (2015) Single sample face recognition via learning deep supervised autoencoders. IEEE Trans Inf Forensics Secur 10(10):2108–2118
Gao X, Wang N, Tao D, Li X (2012) Face sketch-photo synthesis and retrieval using sparse representation. IEEE Trans Circuits Syst Video Technol 22(8):1213–1226. https://doi.org/10.1109/TCSVT.2012.2198090
Gao Y, Ma J, Yuille AL (2017) Semi-supervised sparse representation based classification for face recognition with insufficient labeled samples. IEEE Trans Image Process 26(5):2545–2560
Garrido P, Valgaerts L, Rehmsen O, Thormahlen T, Perez P, Theobalt C (2014) Automatic face reenactment. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4217–4224
Georghiades AS, Belhumeur PN, Kriegman DJ (2001) From few to many: illumination cone models for face recognition under variable lighting and pose. IEEE Trans Pattern Anal Mach Intell 23(6):643–660
Gong D, Li Z, Huang W, Li X, Tao D (2017) Heterogeneous face recognition: a common encoding feature discriminant approach. IEEE Trans Image Process 26(5):2079–2089. https://doi.org/10.1109/TIP.2017.2651380
Grill J, Strub F, Altché F, Tallec C, Richemond PH, Buchatskaya E, Doersch C, Pires BÁ, Guo Z, Azar MG, Piot B, Kavukcuoglu K, Munos R, Valko M (2020) Bootstrap your own latent: a new approach to self-supervised learning. In: H. Larochelle, M. Ranzato, R. Hadsell, M. Balcan, H. Lin (eds) Advances in neural information processing systems 33: annual conference on neural information processing systems 2020, NeurIPS 2020, December 6–12, 2020, virtual. https://proceedings.neurips.cc/paper/2020/hash/f3ada80d5c4ee70142b17b8192b2958e-Abstract.html
Guo Y, Jiao L, Wang S, Wang S, Liu F (2017) Fuzzy sparse autoencoder framework for single image per person face recognition. IEEE Trans Cybern 48(8):2402–2415
Guo Y, Zhang L (2017) One-shot face recognition by promoting underrepresented classes. arXiv:1707.05574
Guo Y, Zhang L, Hu Y, He X, Gao J (2016) Ms-celeb-1m: a dataset and benchmark for large-scale face recognition. In: European conference on computer vision. Springer, pp 87–102
Han H, Klare BF, Bonnen K, Jain AK (2012) Matching composite sketches to face photos: a component-based approach. IEEE Trans Inf Forensics Secur 8(1):191–204
Hao H, Baireddy S, Reibman AR, Delp EJ (2020) Far-gan for one-shot face reenactment. arXiv:2005.06402
He K, Fan H, Wu Y, Xie S, Girshick RB (2020) Momentum contrast for unsupervised visual representation learning. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020, Seattle, WA, USA, June 13–19, 2020, pp 9726–9735. Computer Vision Foundation / IEEE. https://doi.org/10.1109/CVPR42600.2020.00975
He R, Li Y, Wu X, Song L, Chai Z, Wei X (2021) Coupled adversarial learning for semi-supervised heterogeneous face recognition. Pattern Recogn 110:107618
Hong S, Im W, Ryu J, Yang HS (2017) Sspp-dan: deep domain adaptation network for face recognition with single sample per person. In: 2017 IEEE international conference on image processing (ICIP). IEEE, pp 825–829
Huang GB, Mattar M, Berg T, Learned-Miller E (2008) Labeled faces in the wild: a database forstudying face recognition in unconstrained environments
Huang R, Zhang S, Li T, He R (2017) Beyond face rotation: Global and local perception gan for photorealistic and identity preserving frontal view synthesis. In: Proceedings of the IEEE international conference on computer vision, pp 2439–2448
Jadhav A, Namboodiri VP, Venkatesh K (2016) Deep attributes for one-shot face recognition. In: European conference on computer vision. Springer, pp. 516–523
Ji HK, Sun QS, Ji ZX, Yuan YH, Zhang GQ (2017) Collaborative probabilistic labels for face recognition from single sample per person. Pattern Recogn 62:125–134
Kadam S, Vaidya V (2020) Review and analysis of zero, one and few shot learning approaches. In: Abraham A, Cherukuri AK, Melin P, Gandhi N (eds) Intelligent systems design and applications. Springer, Cham, pp 100–112
Chapter Google Scholar
Kan M, Shan S, Su Y, Xu D, Chen X (2013) Adaptive discriminant learning for face recognition. Pattern Recogn 46(9):2497–2509
Karras T, Aittala M, Laine S, Härkönen E, Hellsten J, Lehtinen J, Aila T (2021) Alias-free generative adversarial networks. CoRR arXiv:abs/2106.12423
Klare B, Jain AK (2013) Heterogeneous face recognition using kernel prototype similarities. IEEE Trans Pattern Anal Mach Intell 35(6):1410–1422. https://doi.org/10.1109/TPAMI.2012.229
Klare B, Li Z, Jain AK (2011) Matching forensic sketches to mug shot photos. IEEE Trans Pattern Anal Mach Intell 33(3):639–646. https://doi.org/10.1109/TPAMI.2010.180
Klum SJ, Han H, Klare BF, Jain AK (2014) The facesketchid system: matching facial composites to mugshots. IEEE Trans Inf Forensics Secur 9(12):2248–2263
Kosarevych I, Petruk M, Kostiv M, Kupyn O, Maksymenko M, Budzan V (2020) Actgan: Flexible and efficient one-shot face reenactment. In: 2020 8th international workshop on biometrics and forensics (IWBF). IEEE, pp 1–6
Kumar N, Garg V (2019) Single sample face recognition in the last decade: a survey. Int J Pattern Recognit Artif Intell 33(13):1956009
Li A, Shan S, Chen X, Gao W (2009) Maximizing intra-individual correlations for face recognition across pose differences. In: 2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2009), 20–25 June 2009, IEEE Computer Society, Miami, Florida, USA, pp 605–611. https://doi.org/10.1109/CVPR.2009.5206659
Li JB, Pan JS, Chu SC (2007) Face recognition from a single image per person using common subfaces method. In: International symposium on neural networks. Springer, pp. 905–912
Li L, Peng Y, Qiu G, Sun Z, Liu S (2018) A survey of virtual sample generation technology for face recognition. Artif Intell Rev 50(1):1–20
Li M, Zuo W, Zhang D (2016) Convolutional network for attribute-driven and identity-preserving human face generation. arXiv:1608.06434
Li X, Song A (2011) Face recognition using m-msd and svd with single training image. In: Proceedings of the 30th Chinese control conference. IEEE, pp 3231–3233
Li Z, Liu F, Yang W, Peng S, Zhou J (2021) A survey of convolutional neural networks: analysis, applications, and prospects. IEEE Trans Neural Netw Learn Syst, pp 1–21. https://doi.org/10.1109/TNNLS.2021.3084827
Liu D, Gao X, Wang N, Li J, Peng C (2020) Coupled attribute learning for heterogeneous face recognition. IEEE Trans Neural Netw Learn Syst 31(11):4699–4712
Liu F, Ding Y, Xu F, Ye Q (2019) Learning low-rank regularized generic representation with block-sparse structure for single sample face recognition. IEEE Access 7:30573–30587
Liu F, Tang J, Song Y, Zhang L, Tang Z (2015) Local structure-based sparse representation for face recognition. ACM Trans Intell Syst Technol (TIST) 7(1):1–20
Liu F, Tang J, Song Y, Zhang L, Tang Z (2016) Local structure based multi-phase collaborative representation for face recognition with single sample per person. Inf Sci, pp 346–347
Liu J, Chen S, Zhou ZH, Tan X (2007) Single image subspace for face recognition. In: International workshop on analysis and modeling of faces and gestures. Springer, pp 205–219
Majumdar A, Ward RK (2008) Single image per person face recognition with images synthesized by non-linear approximation. In: 2008 15th IEEE international conference on image processing. IEEE, pp 2740–2743
Malisiewicz T, Gupta A, Efros AA (2011) Ensemble of exemplar-svms for object detection and beyond. In: 2011 International conference on computer vision. IEEE, pp 89–96
Martinez AM (1998) The ar face database. CVC Technical Report24
Martínez AM (2002) Recognizing imprecisely localized, partially occluded, and expression variant faces from a single sample per class. IEEE Trans Pattern Anal Mach Intell 24(6):748–763
Min R, Xu S, Cui Z (2019) Single-sample face recognition based on feature expansion. IEEE Access 7:45219–45229
Minaee S, Luo P, Lin Z, Bowyer KW (2021) Going deeper into face detection: a survey. arXiv:abs/2103.14983
Mokhayeri F, Granger E (2020) A paired sparse representation model for robust face recognition from a single sample. Pattern Recogn 100:107129
Ouanan H, Ouanan M, Aksasse B (2018) Non-linear dictionary representation of deep features for face recognition from a single sample per person. Procedia Comput Sci 127:114–122
Omahony N, Campbell S, Carvalho A, Krpalkova L, Hernandez GV, Harapanahalli S, Riordan D, Walsh J (2019) One-shot learning for custom identification tasks; a review. Procedia Manuf 38:186–193. https://doi.org/10.1016/j.promfg.2020.01.025
Pang M, Cheung YM, Wang B, Lou J (2019) Synergistic generic learning for face recognition from a contaminated single sample per person. IEEE Trans Inf Forensics Secur 15:195–209
Pang M, Cheung YM, Shi Q, Li M (2020) Iterative dynamic generic learning for face recognition from a contaminated single-sample per person. IEEE transactions on neural networks and learning systems
Pang M, Wang B, Cheung YM, Chen Y, Wen B (2021a) Vd-gan: a unified framework for joint prototype and representation learning from contaminated single sample per person. IEEE Trans Inf Forensics Secur 16:2246–2259
Pang M, Wang B, Ye M, Chen Y, Wen B (2021b) Disentangling prototype and variation for single sample face recognition. In: 2021 IEEE International conference on multimedia and expo (ICME). IEEE, pp 1–6
Parchami M, Bashbaghi S, Granger E (2017a) Cnns with cross-correlation matching for face recognition in video surveillance using a single training sample per person. In: 2017 14th IEEE international conference on advanced video and signal based surveillance (AVSS). IEEE, pp 1–6
Parchami M, Bashbaghi S, Granger E (2017b) Video-based face recognition using ensemble of haar-like deep convolutional neural networks. In: 2017 International joint conference on neural networks (IJCNN). IEEE, pp 4625–4632
Phillips PJ, Wechsler H, Huang J, Rauss PJ (1998) The feret database and evaluation procedure for face-recognition algorithms. Image Vis Comput 16(5):295–306
Pumarola A, Agudo A, Martinez AM, Sanfeliu A, Moreno-Noguer F (2018) Ganimation: natomically-aware facial animation from a single image. In: Proceedings of the European conference on computer vision (ECCV), pp 818–833
Reed S, Sohn K, Zhang Y, Lee H (2014) Learning to disentangle factors of variation with manifold interaction. In: International conference on machine learning, pp 1431–1439
Schroff F, Kalenichenko D, Philbin J (2015) Facenet: a unified embedding for face recognition and clustering. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 815–823
Sharma A, Jacobs DW (2011) Bypassing synthesis: PLS for face recognition with pose, low-resolution and sketch. In: The 24th IEEE conference on computer vision and pattern recognition, CVPR 2011. IEEE Computer Society, Colorado Springs, CO, USA, 20–25 June 2011, pp 593–600. https://doi.org/10.1109/CVPR.2011.5995350
Smirnov E, Melnikov A, Novoselov S, Luckyanets E, Lavrentyeva G (2017) Doppelganger mining for face representation learning. In: Proceedings of the IEEE international conference on computer vision workshops, pp 1916–1923
Su Y, Shan S, Chen X, Gao W (2010) Adaptive generic learning for face recognition from a single sample per person. In: 2010 IEEE computer society conference on computer vision and pattern recognition. IEEE, pp 2699–2706
Tan X, Chen S, Zhou ZH, Zhang F (2006) Face recognition from a single image per person: a survey. Pattern Recogn 39(9):1725–1745
Article MATH Google Scholar
Tang Y, Salakhutdinov R, Hinton G (2012) Deep lambertian networks. arXiv:1206.6445
Tran L, Yin X, Liu X (2018) Representation learning by rotating your faces. IEEE Trans Pattern Anal Mach Intell 41(12):3007–3021
Tu H, Duoji G, Zhao Q, Wu S (2020) Improved single sample per person face recognition via enriching intra-variation and invariant features. Appl Sci 10(2):601
Tuan Tran A, Hassner T, Masi I, Medioni G (2017) Regressing robust and discriminative 3d morphable models with a very deep neural network. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 5163–5172
Vasilescu MAO, Terzopoulos D (2003) Multilinear subspace analysis of image ensembles. In: Proceedings 2003 IEEE Computer Society Conference on Computer Vision and Pattern Recognition. . IEEE, vol 2, pp II–93
Vega PJS, Feitosa RQ, Quirita VHA, Happ PN (2016) Single sample face recognition from video via stacked supervised auto-encoder. In: 2016 29th SIBGRAPI conference on graphics, patterns and images (SIBGRAPI). IEEE, pp. 96–103
Viktorisa OY, Wasito I, Syafiandini AF (2016) Evaluating the performance of deep supervised auto encoder in single sample face recognition problem using kullback-leibler divergence sparsity regularizer. J Theor Appl Inf Technol 87(2):255–258
Google Scholar
Vinyals O, Blundell C, Lillicrap T, Wierstra D et al (2016) Matching networks for one shot learning. Adv Neural Inf Process Syst 29:3630–3638
Wang J, Plataniotis KN, Lu J, Venetsanopoulos AN (2006) On solving the face recognition problem with one training sample per subject. Pattern Recogn 39(9):1746–1762
Wang L, Li Y, Wang S (2018) Feature learning for one-shot face recognition. In: 2018 25th IEEE international conference on image processing (ICIP). IEEE, pp. 2386–2390
Wang M, Deng W (2018) Deep visual domain adaptation: a survey. Neurocomputing 312:135–153
Wang M, Deng W (2021) Deep face recognition: a survey. Neurocomputing 429:215–244. https://doi.org/10.1016/j.neucom.2020.10.081
Wang N, Gao X, Li J (2018) Random sampling for fast face sketch synthesis. Pattern Recognit 76:215–227. https://doi.org/10.1016/j.patcog.2017.11.008
Wang X, Zhang B, Yang M, Ke K, Zheng W (2019) Robust joint representation with triple local feature for face recognition with single sample per person. Knowl-Based Syst 181:104790
Wen W, Wang X, Shen L, Yang M (2018) Adaptive convolution local and global learning for class-level joint representation of face recognition with single sample per person. In: 2018 24th International conference on pattern recognition (ICPR). IEEE, pp 3537–3542
Wolf L, Hassner T, Taigman Y (2009) The one-shot similarity kernel. In: 2009 IEEE 12th international conference on computer vision. IEEE, pp 897–902
Wu X, Huang H, Patel VM, He R, Sun Z (2019) Disentangled variational representation for heterogeneous face recognition. In: Proceedings of the AAAI conference on artificial intelligence, vol 33, pp 9005–9012
Wu Y, Liu H, Fu Y (2017) Low-shot face recognition with hybrid classifiers. In: Proceedings of the IEEE international conference on computer vision workshops, pp 1933–1939
Wu Z, Deng W (2016) One-shot deep neural network for pose and illumination normalization face recognition. In: 2016 IEEE international conference on multimedia and expo (ICME). IEEE, pp. 1–6
Yang M, Van Gool L, Zhang L (2013) Sparse variation dictionary learning for face recognition with a single training sample per person. In: Proceedings of the IEEE international conference on computer vision, pp 689–696
Yang M, Wang X, Zeng G, Shen L (2017) Joint and collaborative representation with local adaptive convolution feature for face recognition with single sample per person. Pattern Recogn 66:117–128
Yang M, Zhang L, Feng X, Zhang D (2011) Fisher discrimination dictionary learning for sparse representation. In: 2011 International conference on computer vision. IEEE, pp 543–550
Yao G, Yuan Y, Shao T, Zhou K (2020) Mesh guided one-shot face reenactment using graph convolutional networks. In: Proceedings of the 28th ACM international conference on multimedia, pp 1773–1781
Yin X, Yu X, Sohn K, Liu X, Chandraker M (2019) Feature transfer learning for face recognition with under-represented data. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 5704–5713
You F, Cao Y, Zhang C (2017) Deep domain adaptation with a few samples for face identification. In: 2017 4th IAPR Asian conference on pattern recognition (ACPR). IEEE, pp 178–183
Yu S, Han H, Shan S, Dantcheva A, Chen X (2019) Improving face sketch recognition via adversarial sketch-photo transformation. In: 2019 14th IEEE international conference on automatic face & gesture recognition (FG 2019). IEEE, pp 1–8
Zakharov E, Shysheya A, Burkov E, Lempitsky V (2019) Few-shot adversarial learning of realistic neural talking head models. In: Proceedings of the ieee international conference on computer vision, pp 9459–9468
Zeng J, Zhao X, Qin C, Lin Z (2017) Single sample per person face recognition based on deep convolutional neural network. In: 2017 3rd IEEE international conference on computer and communications (ICCC). IEEE, pp. 1647–1651
Zhang L, Liu J, Zhang B, Zhang D, Zhu C (2019) Deep cascade model-based face recognition: when deep-layered learning meets small data. IEEE Trans Image Process 29:1016–1029
Zhang W, Wang X, Tang X (2011) Coupled information-theoretic encoding for face photo-sketch recognition. In: The 24th IEEE conference on computer vision and pattern recognition, CVPR 2011. IEEE Computer Society, Colorado Springs, CO, USA, 20–25 June 2011, pp 513–520. https://doi.org/10.1109/CVPR.2011.5995324
Zhang W, Wang X, Tang X (2011) Coupled information-theoretic encoding for face photo-sketch recognition. In: CVPR 2011. IEEE, pp 513–520
Zhang Y, Hu C, Lu X (2019) Il-gan: illumination-invariant representation learning for single sample face recognition. J Vis Commun Image Represent 59:501–513
Zhang Y, Peng H (2017) Sample reconstruction with deep autoencoder for one sample per person face recognition. IET Comput Vis 11(6):471–478
Zhang Y, Zhang S, He Y, Li C, Loy CC, Liu Z (2019) One-shot face reenactment. arXiv:1908.03251
Zhou J, Chen J, Liang C, Chen J (2020) One-shot face recognition with feature rectification via adversarial learning. In: International conference on multimedia modeling. Springer, pp 290–302
Zhu Z, Luo P, Wang X, Tang X (2013) Deep learning identity-preserving face space. In: Proceedings of the IEEE International conference on computer vision, pp 113–120
Zhu Z, Luo P, Wang X, Tang X (2014) Multi-view perceptron: a deep model for learning face identity and view representations. Adv Neural Inf Process Syst 27:217–225
Download references
This work was partially funded by Natural Science Foundation of Jiangsu Province under Grant No. BK20191298, Research Fund from Science and Technology on Underwater Vehicle Technology Laboratory (2021JCJQ-SYSJJ-LB06905), Water Science and Technology Project of Jiangsu Province under Grant Nos. 2021072, 2021063.
Fan Liu and Delong Chen have contributed equally to this work.
College of Computer and Information, Hohai University, Nanjing, China
Fan Liu, Delong Chen, Fei Wang, Zewen Li & Feng Xu
Science and Technology on Underwater Vehicle Technology Laboratory, Harbin Engineering University, Harbin, 150001, China
You can also search for this author in PubMed Google Scholar
Correspondence to Fan Liu .
Conflict of interest.
The authors declare that they have no conflict of interest.
Publisher's note.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Springer Nature or its licensor holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
Reprints and permissions
Liu, F., Chen, D., Wang, F. et al. Deep learning based single sample face recognition: a survey. Artif Intell Rev 56 , 2723–2748 (2023). https://doi.org/10.1007/s10462-022-10240-2
Download citation
Published : 05 August 2022
Issue Date : March 2023
DOI : https://doi.org/10.1007/s10462-022-10240-2
Anyone you share the following link with will be able to read this content:
Sorry, a shareable link is not currently available for this article.
Provided by the Springer Nature SharedIt content-sharing initiative
Thank you for visiting nature.com. You are using a browser version with limited support for CSS. To obtain the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles and JavaScript.
Scientific Reports volume 14 , Article number: 17802 ( 2024 ) Cite this article
454 Accesses
5 Altmetric
Metrics details
The PI20 is a self-report questionnaire that assesses the presence of lifelong face recognition difficulties. The items on this scale ask respondents to assess their face recognition ability relative to the rest of the population, either explicitly or implicitly. Recent reports suggest that the PI20 scores of autistic participants exhibit little or no correlation with their performance on the Cambridge Face Memory Test—a key measure of face recognition ability. These reports are suggestive of a meta-cognitive deficit whereby autistic individuals are unable to infer whether their face recognition is impaired relative to the wider population. In the present study, however, we observed significant correlations between the PI20 scores of 77 autistic adults and their performance on two variants of the Cambridge Face Memory Test. These findings indicate that autistic individuals can infer whether their face recognition ability is impaired. Consistent with previous research, we observed a wide spread of face recognition abilities within our autistic sample. While some individuals approached ceiling levels of performance, others met the prevailing diagnostic criteria for developmental prosopagnosia. This variability showed little or no association with non-verbal intelligence, autism severity, or the presence of co-occurring alexithymia or ADHD.
Introduction.
Historically, lifelong face recognition difficulties were thought to be extremely rare 1 . Over the last twenty years, however, there has been growing appreciation that ‘congenital’ or ‘developmental’ prosopagnosia is far more common than was once believed 2 , 3 , 4 , 5 . Indeed, around 2% of the general population describe lifelong face recognition problems severe enough to disrupt their daily lives 6 , 7 . The incidence of lifelong face recognition difficulties is particularly high amongst autistic individuals, many of whom experience problems when asked to identify or match faces 8 , 9 , 10 , 11 .
Increasing awareness of these difficulties has fuelled the development of tools for the identification and assessment of face recognition impairments. One well-known measure is the Cambridge Face Memory Test (CFMT) 12 , 13 , a standardized objective test of face recognition ability that was developed to identify cases of developmental prosopagnosia. On each trial (72 in total), participants are asked to identify recently learned target faces from a line-up of three options (a target and two foils). The addition of view-point changes and high-spatial frequency visual noise increases task difficulty in the later stages. The CFMT has good internal reliability 12 , 13 and correlates well with other measures of face identification 14 .
A second measure developed to aid the identification of developmental prosopagnosia is the Twenty Item Prosopagnosia Index (PI20) 15 , 16 , 17 . This self-report questionnaire was designed to provide standardized self-report evidence of face recognition difficulties, to complement diagnostic evidence obtained from objective computer-based assessments such as the CFMT. The PI20 comprises 20 statements describing face recognition experiences drawn from qualitative and quantitative descriptions of individuals with lifelong face recognition difficulties. Respondents (typically adults) rate how well each statement describes their own experiences on a 5-point scale. Scores can range from 20 to 100. A score of 65 or higher is thought to indicate the likely presence of face recognition impairment. The PI20, originally written in English, has been translated into multiple languages (e.g., Italian, Portuguese, Danish, Japanese & Mandarin) and applied in various cultural contexts 18 , 19 , 20 , 21 , 22 . The twenty items comprising the PI20 can be viewed in the supplementary materials (Table S1 ).
The items on the PI20 ask respondents to assess their face recognition ability relative to the rest of the population, either explicitly (e.g., My face recognition ability is worse than most people; I am better than most people at putting a ‘name to a face’; I have to try harder than other people to memorize faces) or implicitly (e.g., When people change their hairstyle or wear hats, I have problems recognizing them; when I was at school, I struggled to recognize my classmates). There has been considerable debate about whether participants have the necessary insight into their relative face recognition ability to provide meaningful responses to such items 23 , 24 , 25 , 26 . However, there is now strong evidence that the PI20 scores of non-autistic participants correlate with their performance on objective measures of face recognition accuracy 15 , 16 , 17 , 27 . While participants may lack fine-grained insight into their face recognition ability (e.g., whether they fall within the 45th or 55th percentile), these findings suggest that respondents have enough insight to provide meaningful responses on the PI20; i.e., they appear to have some idea whether their face recognition is impaired or unimpaired.
This may not be true of autistic individuals, however. Minio-Paluello and colleagues 28 reported that the PI20 scores of autistic adults ( N = 63) exhibited little or no correlation with their performance on the CFMT—a key objective test of face recognition ability. A similar result was described by Stantić and colleagues 10 . In this study, the authors observed a non-significant correlation of r = − 0.17 between the PI20 scores of 31 autistic adults and their performance on the CFMT. If found to be robust, these results have important theoretical implications: they raise the possibility that face recognition in autism may be subject to a metacognitive deficit, whereby autistic individuals are unable to infer whether (or not) their face recognition ability is impaired relative to the wider population. There is also an important substantive implication. These results suggests that the PI20 may not be suitable for screening autistic participants for face recognition difficulties. This would be a non-trivial limitation, not least because face recognition difficulties appear to be far more prevalent in the autistic population than in the non-autistic population 8 , 9 , 10 , 11 .
There are several reasons to be cautious when interpreting these findings, however. First, previous research suggests that metacognitive differences in autistic adults tend to be small and subtle, if observed at all 29 . Second, the study described by Stantić et al. 10 was not designed to examine individual differences. Any conclusions about face recognition variability and correlations therewith, are limited by the relatively small size of the study’s autistic sample ( N = 31). Correlation estimates obtained with small samples are notoriously unstable 30 . Third, both results were obtained using the original version of the CFMT (the CFMT-O) 13 . This version of the CFMT is now easily accessible online; it is hosted by several websites, and various prosopagnosia forums and pop-science resources link to this test. Consequently, many individuals with face recognition difficulties have attempted the CFMT-O on multiple occasions 31 . Where practice benefits arise, participants may achieve higher scores than might be expected based on their PI20 score.
In light of the foregoing observations, we were keen to re-examine the relationship between the PI20 scores and CFMT performance of autistic individuals. To this end, a group of 77 autistic participants completed the PI20 questionnaire and two variants of the CFMT: the original (CFMT-O) 13 and the Australian (CFMT-A) 12 versions. The CFMT-O and CFMT-A share an identical format and differ only in terms of the (White male) facial identities used. Unlike the CFMT-O, however, the CFMT-A is not widely available to the members of the general public.
It has been noted previously that the face recognition abilities of autistic participants vary widely 8 , 9 , 10 , 11 , 28 . At present, however, little is known about the nature and origin of this variability. Some of this variance might be explained by differences in autism severity 32 . However, performance on face processing tasks may also be affected by differences in non-verbal intelligence 33 and the presence of co-occurring conditions, notably alexithymia 34 , 35 and attention-deficit-hyperactivity disorder (ADHD) 36 , 37 . We therefore took this opportunity to explore which of these factors—if any—predicted face recognition performance in our autistic sample.
Seventy-seven participants with a clinical diagnosis of autism ( M age = 35.99 years; SD a ge = 11.60 years) were recruited via www.ukautismresearch.org . All autistic participants had received an autism diagnosis (e.g., Autism Spectrum Disorder, Asperger’s Syndrome) from a clinical professional (General Practitioner, Neurologist, Psychiatrist or Clinical Psychologist) based in the U.K. All participants in the autistic group also reached cut-off (a score of 32) on the Autism Spectrum Quotient (AQ) 38 . The mean AQ score of the participants was 42.45 ( SD = 4.17). To be eligible, participants also had to be aged between 18 and 60, speak English as a first language and have normal or corrected-to-normal visual acuity. All participants were required to be a current U.K. resident.
Of the 16 individuals who described their sex as male, 13 described their gender identity as male, 2 identified as non-binary and 1 identified as female. Of the 61 individuals who described their sex as female, 48 described their gender identity as female, 9 identified as non-binary, 1 as male and 3 preferred not to say. Seventy-six of the 77 participants identified as White (73: White-British, 1: White Irish, 2: White-Other). One participant did not specify their ethnicity. Sixty-eight of the participants were right-handed, while 9 were left-handed.
Data collection for the study took place between June and August 2023. At the outset, our aim was i) to recruit as many participants as possible during this period, and ii) to stop data collection at the end of August provided a minimum sample size of N = 62 had been reached. A sample of N = 62 yields a 90% chance of detecting a correlation of r = 0.40 between PI20 scores and CFMT performance. Our final sample ( N = 77) comfortably exceeded this minimum.
Ethical clearance was granted by the Departmental Ethics Committee for Psychological Sciences, Birkbeck, University of London and the experiment was conducted in line with the ethical guidelines laid down in the 6th (2008) Declaration of Helsinki. All participants gave informed consent before taking part.
The principal aim of the study was to elucidate the relationship between participants’ PI20 scores and their performance on the CFMT. To this end, all participants completed the PI20 questionnaire 16 and two versions of the CFMT: the CFMT-O 13 and the CFMT-A 12 . All participants completed the PI20 before attempting the CFMTs. Participants also completed the AQ to confirm their eligibility for the study. In addition to these measures, all participants completed a self-report measure of alexithymia severity: the Twenty-item Toronto Alexithymia Scale (TAS20) 39 , 40 , a self-report measure of ADHD traits: the Adult ADHD Self-Report Scale (ASRS) 41 , and a matrix reasoning task (MRT) to assess their non-verbal intelligence.
The TAS20 comprises 20 statements that relate to one’s ability to describe and identify emotions and interoceptive sensations. Respondents indicate to what extent each statement applies to them on a 5-point scale. Scores can range from 20 to 100, with higher scores indicative of more alexithymic traits. A score of 61 or higher is thought to index clinically significant levels of alexithymia. The TAS20 has good psychometric properties 39 and is widely used to assess the presence of alexithymia in autistic and non-autistic individuals 34 .
The ASRS is a self-report questionnaire that assesses the presence of traits associated with inattention, hyperactivity, and impulsivity. The ASRS consists of two parts: Part A is a 6-item screener that has been shown to effectively discriminate clinical cases of adult ADHD from non-cases 42 . Each response is scored as either 0 or 1, thus screener scores can range from 0 to 6. A score of 4 or above is thought to be associated with clinically significant levels of ADHD traits. Part B consists of 12 follow-up items that can be used to probe symptomology. Part B was not employed here.
The MRT employed consists of forty items selected from The Matrix Reasoning Item Bank 43 . Participants were given 30 s to complete each puzzle by selecting the correct answer from 4 options. Participants responded using keyboard number keys (1–4), were given a 5-s warning before the end of each trial, and received no feedback. Each participant attempted all forty items. Participants had to complete 3 practice trials correctly before beginning the test. We have employed this measure in previous studies of social perception in autism 8 , 35 . Based on a sample of 100 non-autistic adults ( M age = 34.90; SD a ge = 10.16), we estimate the test–retest reliability of this measure to be r p = 0.727 (see Supplementary Material ). All data reported here were collected online. Both versions of the CFMT and the matrix reasoning test were administered via Gorilla Experiment Builder 44 .
The correlational analyses described below (all α = 0.05, two-tailed) were conducted using Pearson’s r ( r p ) and Spearman’s rho ( r s ) . The comparison of autistic subgroups was assessed using independent samples t -tests (α = 0.05, two-tailed). For each t -test we also provide the associated Bayes factor (BF), calculated in JASP 45 with default prior width. We interpret BFs of less than 3.0 as anecdotal evidence for the null hypothesis. BFs of greater than 3.0 are treated as substantial evidence for the null hypothesis 46 . The data supporting all the analyses described are available via the Open Science Framework ( https://osf.io/tesk5/ ).
The mean scores obtained for each measure are shown in Table 1 . As expected, we saw strong correlation between performance on the CFMT-O and CFMT-A [ N = 77, r p = 0.744, p < 0.001], underscoring the good psychometric properties of our two dependent measures. We also observed a number of significant correlations between our predictor variables (Table 1 ). Reassuringly, several of these relationships are predicted by the existing literature 34 , 47 , 48 , including the AQ-TAS20 correlation [ N = 77, r p = 0.526, p < 0.001] and the AQ-ASRS correlation [ N = 77, r p = 0.322, p = 0.004]. The mean PI20 score ( M = 62.43) accords well with the mean PI20 score described by Stantić and colleagues 10 ( M = 63.30) but is a little higher than that reported by Minio-Paluello and colleagues 28 ( M = 55.5).
The present study had two principal aims: first, to establish whether or not the PI20 scores of autistic adults correlate with their CFMT performance. Second, to explore whether differences in non-verbal intelligence and the presence of co-occurring conditions (alexithymia and ADHD) account for the enormous variability in face recognition ability seen in the autistic population. Thus, the focus of our investigation is the variability in face recognition performance observed within the autistic sample.
At the outset of our analyses, however, we first sought to evaluate the overall performance of the autistic sample on the CFMT-O ( M = 67.95, SD = 15.80) and CFMT-A ( M = 70.67, SD = 15.22). For this purpose, we employed comparison data reported by Tsantani et al. 17 (Fig. 1 a). These data were obtained from 238 non-autistic individuals (131 females, 104 males, 3 non-binary; M age = 36.56, SD age = 11.72), who completed online versions of the CFMT-O ( M = 73.96, SD = 13.77) and CFMT-A ( M = 75.37, SD = 12.48) under similar conditions. The participants in this sample were recruited via Prolific ( www.prolific.com ). Thirteen of the 238 participants (5.46%) reached the PI20 cut-off score of 65 ( M = 44.85, SD = 10.70).
( a ) Mean scores on the CFMT-O and CFMT-A for the autistic sample. ( b ) Mean scores on the CFMT-O and CFMT-A for those autistic participants who reached the PI20 cut-off score (high-scorers) and those who did not (low-scorers). The non-autistic comparison data illustrated in both panels is taken from Tsantani et al. 17 . ** p ≤ 0.01, *** p ≤ 0.001. Error bars denote ± 1SD.
As expected, the scores of the autistic participants in our sample were significantly below those seen in this comparison sample, both for the CFMT-O [ t (313) = 3.207, d = 0.420, p = 0.001, BF 01 = 0.057] and the CFMT-A [ t (110.97) = 2.453, p = 0.016, d = 0.356, BF 01 = 0.221]. Note, for this latter comparison it was necessary to correct the degrees of freedom because the variance in our autistic sample was greater than that seen in the non-autistic comparison data [ F (1, 313) = 5.387, p = 0.021]. The fact that the CFMT scores of our autistic sample tended to be lower than those of the non-autistic comparison sample accords well the existing literature 8 , 9 , 10 , 11 . This finding suggests that, in terms of face recognition ability, our autistic sample is broadly comparable with autistic samples described elsewhere.
Next, we sought to determine whether the PI20 scores of our autistic participants were predictive of their CFMT performance. To begin, we examined the simple correlations between participants’ PI20 and CFMT scores. Contrary to the findings of Minio-Paluello et al. 28 and Stantić et al. 10 , we observed significant correlation between PI20 scores and performance on both the CFMT-O [ N = 77, r p = − 0.486, p < 0.001] and CFMT-A [ N = 77, r p = − 0.464, p < 0.001] (Fig. 2 ). Similar correlations were seen when the raw scores were transformed into ranks for both the CFMT-O [ N = 77, r s = − 0.435, p < 0.001] and CFMT-A [ N = 77, r s = − 0.469, p < 0.001].
Scatterplots of the relationship between PI20 scores and performance on the CFMT-O (left) and the CFMT-A (right). The solid lines depict the linear trends. The dashed lines depict the mean performance of the non-autistic sample described by Tsantani et al. 17 .
We also conducted a complementary subgroup analysis based on the established PI20 cut-off of 65. We split our sample of 77 autistic participants into those who met the cut-off ( N = 42, M age = 37.40, SD age = 11.61) and those who did not ( N = 35, M age = 34.29, SD age = 11.51). The autistic participants who met the PI20 cut-off achieved significantly lower scores than those who did not on both the CFMT-O [low-scorers: M = 76.23, SD = 13.59; high-scorers: M = 61.04, SD = 14.21; t (75) = 4.762, p < 0.001, d = 1.090, BF 01 < 0.001] and the CFMT-A [low-scorers: M = 77.54, SD = 12.60; high-scorers: M = 64.95, SD = 14.97; t (75) = 3.946, p < 0.001, d = 0.903, BF 01 = 0.007] (Fig. 1 b). Moreover, the autistic participants who met the PI20 cut-off performed worse on the CFMT-O [ t (278) = 5.576, p < 0.001, d = 0.933, BF 01 < 0.001] and CFMT-A [ t (278) = 4.835, p < 0.001, d = 0.809, BF 01 < 0.001] than the non-autistic participants tested by Tsantani and colleagues 17 (Fig. 1 b). In contrast, the CFMT-O scores [ t (271) = − 0.914, p = 0.362, d = − 0.165, BF 01 = 3.556] and CFMT-A scores [ t (271) = − 0.960, p = 0.338, d = − 0.174, BF 01 = 3.419] of the autistic participants who did not meet the PI20 cut-off did not differ significantly from the comparison distributions described by Tsantani et al. 17 .
There was some correlation between scores on the TAS20—a measure of alexithymia—and performance on the CFMT-A [ N = 77, r p = − 0.252, p = 0.027]. We also observed a significant correlation between TAS20 scores and average performance on the CFMT-O and CFMT-A [ N = 77, r p = − 0.229, p = 0.045]. However, we failed to observe a significant relationship with CFMT-O scores independently [ N = 77, r p = − 0.177, p = 0.125]. We also note that the significant TAS20-CFMT correlations described above do not survive correction for multiple comparisons. No significant correlation was observed between scores on the ASRS—a measure of ADHD traits—and either CFMT-O scores [ N = 77, r p = 0.077, p = 0.504] or CFMT-A scores [ N = 77, r p = 0.064, p = 0.580]. Interestingly, we observed a noteworthy correlation between TAS20 and ASRS scores [ N = 77, r p = 0.453, p < 0.001]; i.e., those autistic participants who reported high levels of alexithymic traits also reported higher levels of ADHD traits (Fig. 3 ).
Simple correlations observed between autism severity (inferred from scores on the AQ questionnaire), levels of alexithymia (inferred from TAS20 scores), and the presence of ADHD traits (inferred from the ASRS screener). All correlations are significant at p < 0.001 ( N = 77).
Like the PI20, both the TAS20 and the ASRS have established cut-offs, associated with clinically significant levels of alexithymia and ADHD traits, respectively. We therefore examined whether subgroup analyses of TAS20 and ASRS scores would reveal evidence of a predictive relationship with CFMT. Of the 77 autistic participants, 59 met the TAS20 cut-off for clinically significant levels of alexithymia, while 18 did not. Those who met cut-off and those who did not, did not differ in their scores on the CFMT-O [low-scorers: M = 70.22, SD = 14.78; high-scorers: M = 67.26, SD = 16.15; t (75) = 0.694, p = 0.490, d = 0.187, BF 01 = 3.016] or on the CFMT-A [low-scorers: M = 74.15, SD = 11.32; high-scorers: M = 69.61, SD = 16.60; t (75) = 1.110, p = 0.271, d = 0.299, BF 01 = 2.212]. Similarly, 51 autistic participants met the ASRS cut-off for clinically significant ADHD traits, while 26 did not. Once again, there was little sign that CFMT-O scores [low-scorers: M = 66.61, SD = 18.64; high-scorers: M = 68.63, SD = 14.29; t (75) = 0.527, p = 0.600, d = 0.127, BF 01 = 3.586] or CFMT-A scores [low-scorers: M = 71.53, SD = 16.29; high-scorers: M = 70.23, SD = 14.80; t (75) = 0.351, p = 0.727, d = 0.084, BF 01 = 3.831] differed across these subgroups.
No significant correlation was observed between AQ scores and CFMT-O scores [ N = 77, r p = − 0.190, p = 0.098] or between AQ scores and CFMT-A scores [ N = 77, r p = − 0.173, p = 0.131]. Note, however, meeting the AQ cut-off score was part of the study inclusion criteria; hence, all 77 autistic participants had an AQ score of 33 or higher. Similarly, no significant correlation was observed between MRT scores and CFMT-O scores [ N = 77, r p = − 0.090, p = 0.436] or between MRT scores and CFMT-A scores [ N = 77, r p = 0.001, p = 0.992]. In sum, we find no evidence in our data that non-verbal intelligence or autism severity influence the face recognition abilities of autistic participants.
There is now considerable evidence that the PI20 scores of non-autistic participants correlate with their performance on objective measures of face recognition accuracy 15 , 16 , 17 , 27 . These findings suggest that respondents have enough insight into their relative face recognition ability to provide meaningful responses on the PI20. Recently, however, Minio-Paluello et al. 28 reported that the PI20 scores of autistic participants ( N = 66) exhibited little or no correlation with their performance on the CFMT. A similar finding was described by Stantić et al. 10 , albeit with a smaller sample ( N = 31). These reports are potentially important because they suggest the possibility that autistic individuals may experience a metacognitive deficit, whereby they are unable to infer whether (or not) their face recognition ability is impaired. Moreover, these results raise the possibility that the PI20 may be unsuitable for screening autistic participants for face recognition difficulties.
Contrary to these reports, however, we find clear evidence of association between the PI20 scores of autistic participants ( N = 77) and their performance on the CFMT-O and the CFMT-A. This association was evident both in simple correlation analyses, and in subgroup analyses where the autistic sample was split into those who met the established cut-off for developmental prosopagnosia, and those who did not. The mean score of those autistic participants who met cut-off was ~ 15% and ~ 12.5% lower than those that did not, on the CFMT-O and CFMT-A, respectively. Indeed, those autistic participants who did not meet the PI20 cut-off exhibited similar levels of performance to a non-autistic comparison sample described previously 17 . Together, these analyses provide clear evidence that the PI20 scores of autistic participants predict their CFMT performance.
The most likely explanation for the failure of Stantić et al. 10 to detect a correlation between scores on the PI20 and the CFMT is the relatively small size of their autistic group ( N = 31). As we allude to in the introduction, (1) this study was not designed to examine the individual differences seen within the autistic population, and (2) correlation estimates obtained with small samples are notoriously unstable 30 . Post-hoc power analysis indicates there is a 38% chance of failing to detect a significant correlation of r = 0.40 with a sample of this size (α = 0.05, two-tailed).
Assuming the authors scored the PI20 correctly, the null correlation described by Minio-Paluello et al. 28 is harder to explain. One relevant factor may be the wide range of general cognitive abilities present in their autistic sample ( N = 63). As a self-report scale, the PI20 has relatively high verbal demands potentially making it unsuitable for individuals with intellectual disability. Moreover, five of the twenty items are reverse scored. Respondents must therefore read individual items carefully to respond appropriately. If some of the participants tested by Minio-Paluello et al. 28 struggled to interpret scale items, and were unable to respond appropriately, this might also explain why the mean PI20 score was lower than that reported here and elsewhere 10 .
It is now beyond doubt that the face recognition abilities of autistic participants vary enormously 8 , 9 , 10 , 11 , 28 . Once again, we saw evidence of this variability in our sample. On the one hand, 13 of our 77 autistic participants (16.9%) scored 65 or higher on the PI20 and scored less than 60% on both versions of the CFMT. These individuals would meet the diagnostic criteria for developmental prosopagnosia employed by the vast majority of research groups 13 , 49 . On the other hand, 10 of the 77 autistic participants (13.0%) scored 85% or higher on both tests, suggestive of excellent face recognition 12 , 17 .
There was little sign in our data that variability in face recognition ability is attributable to differences in non-verbal intelligence (as measured by MRT score), autism severity (as measured by AQ score), or the presence co-occurring ADHD traits (as measured by ASRS score). There was some hint of a potential relationship between the presence of co-occurring alexithymia and face recognition ability: TAS20 scores were negatively correlated with performance on the CFMT-A and with average CFMT performance. However, TAS20 scores did not exhibit significant correlation with CFMT-O scores independently, and the foregoing correlations do not survive correction for multiple-comparisons.
What should we make of this variability? We favour the view that, like alexithymia and ADHD, developmental prosopagnosia is a neurodevelopmental condition that can occur independently of autism, but that also frequently co-occurs with autism 4 , 8 , 51 , 52 . This view not only accounts for the severe lifelong face recognition problems seen in some autistic individuals, but also explains why many other autistic individuals exhibit excellent face recognition. Moreover, this account accords with the prevailing view that the co-occurrence of neurodevelopmental conditions is the ‘norm’ rather than the exception 34 , 47 , 48 , 53 , 54 , 55 . Given what we know about neurodevelopmental conditions more broadly, it would be hugely surprising if the incidence of developmental prosopagnosia was not elevated in the autistic population.
Recently, some authors have rejected this account citing evidence that autistic samples still exhibit below-average face recognition when those who meet the diagnostic criteria for prosopagnosia are removed 11 . However, this critique overlooks two issues. First, diagnostic assessments for developmental prosopagnosia are imperfect 26 . Many autistic individuals with severe co-occurring prosopagnosia may fail to meet diagnostic thresholds simply because of measurement error. Second, the severity of developmental prosopagnosia is thought to vary 56 . While some autistic individuals may experience severe developmental prosopagnosia, others may experience relatively mild forms. These latter individuals may fail to meet conservative diagnostic criteria for developmental prosopagnosia, but still exhibit below average face recognition.
While it was not the focus of our study, we observed a striking correlation between the presence of alexithymia and ADHD traits in our autistic participants. The fact that those autistic participants who report high levels of alexithymia also tend to report high levels of ADHD traits is potentially significant for understanding socio-cognitive differences in autism. In recent years, there has been increasing suggestion that many social perception difficulties traditionally attributed to autism—such as atypical interpretation of facial expressions 35 , 57 and reduced eye-region fixations 50 , 58 —may actually be products of co-occurring alexithymia. Likewise, there is some suggestion that other socio-cognitive differences attributed to autism—for example, atypical attentional cueing by gaze direction 37 —may be partly attributable to co-occurring ADHD. To date, however, authors have tended to assess the presence of either co-occurring alexithymia or co-occurring ADHD. In future, it may prove valuable to establish the extent to which these constructs exert independent or interactive effects in these domains.
We note that 61 of our 77 participants described their sex as female. Conversely, the majority of the autistic population are thought to identify as male 59 . This is not the first time that a high proportion of female participants has been seen where studies have sought to recruit autistic participants online 60 . Unlike participant age 61 , the sex/gender of observers is not thought to exert a strong influence on face recognition ability 62 . However, we acknowledge the need to replicate the present findings in a sample more representative of the wider autistic community.
It is important that future research ascertain if/how other measures of meta-cognitive performance—such as estimates of meta c and meta d inferred from type-II signal detection tasks 63 , 64 —relate to participants’ responses on the PI20. For example, one might hypothesize that the PI20 scores of those with a higher meta c ought to correspond more closely to objective face recognition performance. It might also be interesting to examine how autistic and non-autistic individuals acquire insight into their relative face recognition abilities (e.g., What kinds of face recognition errors are salient? What have individuals been told about face recognition in autism?).
Contrary to recent reports, we observed significant correlation between PI20 scores and performance on both the CFMT-O and CFMT-A in autistic adults. This finding indicates that autistic individuals are able to infer whether (or not) their face recognition ability is impaired and confirms that the PI20 can be used to screen autistic participants for face recognition difficulties. Consistent with previous research, the face recognition performance within our autistic sample varied considerably. While some individuals approached ceiling levels of recognition accuracy, others met the prevailing diagnostic criteria for developmental prosopagnosia. This variability showed little or no association with non-verbal intelligence, autism severity, or the presence of co-occurring alexithymia or ADHD.
The data supporting all the analyses is available here: https://osf.io/tesk5/ .
McConachie, H. R. Developmental prosopagnosia. A single case report. Cortex 12 , 76–82 (1976).
Article CAS PubMed Google Scholar
Wilmer, J. B. Individual differences in face recognition: A decade of discovery. Curr. Dir. Psychol. Sci. 26 , 225–230 (2017).
Article Google Scholar
Behrmann, M. & Avidan, G. Congenital prosopagnosia: Face-blind from birth. Trends Cogn. Sci. 9 , 180–187 (2005).
Article PubMed Google Scholar
Cook, R. & Biotti, F. Developmental prosopagnosia. Curr. Biol. 26 , R312–R313 (2016).
Duchaine, B. & Nakayama, K. Developmental prosopagnosia: A window to content-specific face processing. Curr. Opin. Neurobiol. 16 , 166–173 (2006).
Kennerknecht, I. et al. First report of prevalence of non-syndromic hereditary prosopagnosia (HPA). Am. J. Med. Genet. 140A , 1617–1622 (2006).
Kennerknecht, I., Ho, N. Y. & Wong, V. C. N. Prevalence of heriditary prosopagonsia (HPA) in Hong Kong Chinese population. Am. J. Med. Genet. 146A , 2863–2870 (2008).
Gehdu, B. K., Gray, K. L. & Cook, R. Impaired grouping of ambient facial images in autism. Sci. Rep. 12 , e6665 (2022).
Article ADS Google Scholar
Hedley, D., Brewer, N. & Young, R. Face recognition performance of individuals with Asperger syndrome on the Cambridge Face Memory Test. Autism Res. 4 , 449–455 (2011).
Stantić, M., Ichijo, E., Catmur, C. & Bird, G. Face memory and face perception in autism. Autism , 13623613211027685 (2021).
Kamensek, T., Susilo, T., Iarocci, G. & Oruc, I. Are people with autism prosopagnosic? Autism Res . (2023).
McKone, E. et al. Face ethnicity and measurement reliability affect face recognition performance in developmental prosopagnosia: Evidence from the Cambridge face memory test-Australian. Cogn. Neuropsychol. 28 , 109–146 (2011).
Duchaine, B. & Nakayama, K. The Cambridge Face Memory Test: results for neurologically intact individuals and an investigation of its validity using inverted face stimuli and prosopagnosic participants. Neuropsychologia 44 , 576–585 (2006).
Biotti, F., Gray, K. L. H. & Cook, R. Is developmental prosopagnosia best characterised as an apperceptive or mnemonic condition?. Neuropsychologia 124 , 285–298 (2019).
Gray, K. L. H., Bird, G. & Cook, R. Robust associations between the 20-item prosopagnosia index and the Cambridge Face Memory Test in the general population. Royal Society Open Sci. 4 , 160923 (2017).
Shah, P., Gaule, A., Sowden, S., Bird, G. & Cook, R. The 20-item prosopagnosia index (PI20): A self-report instrument for identifying developmental prosopagnosia. Royal Society Open Sci. 2 , 140343 (2015).
Tsantani, M., Vestner, T. & Cook, R. The Twenty Item Prosopagnosia Index (PI20) provides meaningful evidence of face recognition impairment. Royal Society Open Sci. 8 , e202062 (2021).
Estudillo, A. J. & Wong, H. K. Associations between self-reported and objective face recognition abilities are only evident in above-and below-average recognisers. PeerJ 9 , e10629 (2021).
Article PubMed PubMed Central Google Scholar
Tagliente, S. et al. Self-reported face recognition abilities moderately predict face-learning skills: Evidence from Italian samples. Heliyon 9 , e14125 (2023).
Nørkær, E. et al. The Danish version of the 20-Item prosopagnosia index (PI20): Translation, validation and a link to face perception. Brain Sciences 13 , e337 (2023).
Oishi, Y., Aruga, K. & Kurita, K. (2024), Relationship between face recognition ability and anxiety tendencies in healthy young individuals: A prosopagnosia index and state-trait anxiety inventory study. Acta Psychol. 245 , e104237 (2024).
Ventura, P., Livingston, L. A. & Shah, P. Adults have moderate-to-good insight into their face recognition ability: Further validation of the 20-item Prosopagnosia Index in a Portuguese sample. Quart. J. Exp. Psychol. 71 , 2677–2679 (2018).
Bobak, A. K., Mileva, V. R. & Hancock, P. J. Facing the facts: Naive participants have only moderate insight into their face recognition and face perception abilities. Quart. J. Exp. Psychol. 72 , 872–881 (2019).
Matsuyoshi, D. & Watanabe, K. People have modest, not good, insight into their face recognition ability: A comparison between self-report questionnaires. Psychol. Res. 85 , 1713–1723 (2021).
Arizpe, J. M. et al. Self-reported face recognition is highly valid, but alone is not highly discriminative of prosopagnosia-level performance on objective assessments. Behav. Res. Methods 51 , 1102–1116 (2019).
Burns, E. J., Gaunt, E., Kidane, B., Hunter, L. & Pulford, J. A new approach to diagnosing and researching developmental prosopagnosia: Excluded cases are impaired too. Behav. Res. Methods. https://doi.org/10.3758/s13428-022-02017-w (2022).
Shah, P., Sowden, S., Gaule, A., Catmur, C. & Bird, G. The 20 item prosopagnosia index (PI20): Relationship with the Glasgow face-matching test. Royal Society Open Sci. 2 , e150305 (2015).
Minio-Paluello, I., Porciello, G., Pascual-Leone, A. & Baron-Cohen, S. Face individual identity recognition: A potential endophenotype in autism. Mol. Autism 11 , 1–16 (2020).
Carpenter, K. L. & Williams, D. M. A meta-analysis and critical review of metacognitive accuracy in autism. Autism 27 , 512–525 (2023).
Schönbrodt, F. D. & Perugini, M. At what sample size do correlations stabilize?. J. Res. Personality 47 , 609–612 (2013).
Murray, E. & Bate, S. Diagnosing developmental prosopagnosia: Repeat assessment using the Cambridge Face Memory Test. Royal Society Open Sci. 7 , e200884 (2020).
Keating, C. T., Fraser, D. S., Sowden, S. & Cook, J. L. Differences between autistic and non-autistic adults in the recognition of anger from facial motion remain after controlling for alexithymia. J. Autism Develop. Disord. 52 , 1855–1871 (2022).
Walker, D. L., Palermo, R., Callis, Z. & Gignac, G. E. The association between intelligence and face processing abilities: A conceptual and meta-analytic review. Intelligence 96 , e101718 (2023).
Bird, G. & Cook, R. Mixed emotions: The contribution of alexithymia to the emotional symptoms of autism. Transl. Psychiatry 3 , e285 (2013).
Article CAS PubMed PubMed Central Google Scholar
Gehdu, B. K., Tsantani, M., Press, C., Gray, K. L. & Cook, R. Recognition of facial expressions in autism: Effects of face masks and alexithymia. Quart. J. Exp. Psychol . e17470218231163007 (2023).
Thoma, P., Soria Bauser, D., Edel, M. A., Juckel, G. & Suchan, B. Configural processing of emotional bodies and faces in patients with attention deficit hyperactivity disorder. J. Clin. Exp. Neuropsychol. 42 , 1028–1048 (2020).
Seernani, D. et al. Social and non-social gaze cueing in autism spectrum disorder, attention-deficit/hyperactivity disorder and a comorbid group. Biol. Psychol. 162 , e108096 (2021).
Baron-Cohen, S., Wheelwright, S., Skinner, R., Martin, J. & Clubley, E. The autism-spectrum quotient (AQ): Evidence from asperger syndrome/high-functioning autism, malesand females, scientists and mathematicians. J. Autism Develop. Disorders 31 , 5–17 (2001).
Article CAS Google Scholar
Bagby, R. M., Parker, J. D. & Taylor, G. J. The twenty-item Toronto Alexithymia Scale-I. Item selection and cross-validation of the factor structure. J. Psychosomatic Res. 38 , 23–32 (1994).
Taylor, G. J., Bagby, R. M. & Parker, J. D. The 20-Item Toronto Alexithymia Scale: IV. Reliability and factorial validity in different languages and cultures. J. Psychosomatic Res. 55 , 277–283 (2003).
Kessler, R. C. et al. The World Health Organization Adult ADHD Self-Report Scale (ASRS): A short screening scale for use in the general population. Psychol. Med. 35 , 245–256 (2005).
Kessler, R. C. et al. Validity of the World Health Organization Adult ADHD Self-Report Scale (ASRS) Screener in a representative sample of health plan members. Int. J. Methods Psychiatric Res. 16 , 52–65 (2007).
Chierchia, G. et al. The matrix reasoning item bank (MaRs-IB): Novel, open-access abstract reasoning items for adolescents and adults. Royal Society Open Sci. 6 , 190232 (2019).
Anwyl-Irvine, A. L., Massonnié, J., Flitton, A., Kirkham, N. & Evershed, J. K. Gorilla in our midst: An online behavioral experiment builder. Behav. Res. Methods 52 , 388–407 (2020).
JASP-Team. JASP (Version 0.16.3)[Computer software]. Amsterdam, The Netherlands. (2022).
Jeffreys, H. Theory of probability (3rd ed.) . (Oxford University Press, 1961).
Hours, C., Recasens, C. & Baleyte, J. M. ASD and ADHD comorbidity: What are we talking about?. Front. Psychiatry 13 , e154 (2022).
Leitner, Y. The co-occurrence of autism and attention deficit hyperactivity disorder in children–what do we know? Front. Human Neurosci . 8 (2014).
Tsantani, M., Gray, K. L. H. & Cook, R. New evidence of impaired expression recognition in developmental prosopagnosia. Cortex (2022).
Cuve, H. C. et al. Alexithymia explains atypical spatiotemporal dynamics of eye gaze in autism. Cognition 212 (2021).
Gray, K. L. H. & Cook, R. Should developmental prosopagnosia, developmental body agnosia, and developmental object agnosia be considered independent neurodevelopmental conditions?. Cognit. Neuropsychol. 35 , 59–62 (2018).
Kracke, I. Developmental prosopagnosia in Asperger syndrome: Presentation and discussion of an individual case. Develop. Med. Child Neurol. 36 , 873–886 (1994).
Conti-Ramsden, G., Simkin, Z. & Botting, N. The prevalence of autistic spectrum disorders in adolescents with a history of specific language impairment (SLI). J. Child Psychol. Psychiatry 47 , 621–628 (2006).
Dziuk, M. A. et al. Dyspraxia in autism: Association with motor, social, and communicative deficits. Develop. Med. Child Neurol. 49 , 734–739 (2007).
Gilger, J. W. & Kaplan, B. J. Atypical brain development: A conceptual framework for understanding developmental learning disabilities. Develop. Neuropsychol. 20 , 465–481 (2001).
DeGutis, J. et al. What is the prevalence of developmental prosopagnosia? An empirical assessment of different diagnostic cutoffs. Cortex 161 , 51–64 (2023).
Cook, R., Brewer, R., Shah, P. & Bird, G. Alexithymia, not autism, predicts poor recognition of emotional facial expressions. Psychol. Sci. 24 , 723–732 (2013).
Bird, G., Press, C. & Richardson, D. C. The role of alexithymia in reduced eye-fixation in autism spectrum conditions. J. Autism Develop. Disorders 41 , 1556–1564 (2011).
Ferri, S. L., Abel, T. & Brodkin, E. S. Sex differences in autism spectrum disorder: A review. Curr. Psychiatry Rep. 20 , 1–17 (2018).
Rødgaard, E. M., Jensen, K., Miskowiak, K. W. & Mottron, L. Representativeness of autistic samples in studies recruiting through social media. Autism Res. 15 , 1447–1456 (2022).
Germine, L., Duchaine, B. & Nakayama, K. Where cognitive development and aging meet: Face learning ability peaks after age 30. Cognition 118 , 201–210 (2011).
Gray, K. L. H., Biotti, F. & Cook, R. Evaluating object recognition ability in developmental prosopagnosia using the Cambridge Car Memory Test. Cognitive Neuropsychol. 36 , 89–96 (2019).
Maniscalco, B. & Lau, H. A signal detection theoretic approach for estimating metacognitive sensitivity from confidence ratings. Consciousness Cognition 21 , 422–430 (2012).
Fleming, S. M. & Lau, H. C. How to measure metacognition. Front. Human Neurosci. 8 , e443 (2014).
Download references
Authors and affiliations.
Department of Psychological Sciences, Birkbeck, University of London, London, UK
Bayparvah Kaur Gehdu
Department of Experimental Psychology, University College London, London, UK
Clare Press
Wellcome Centre for Human Neuroimaging, University College London, London, UK
School of Psychology and Clinical Language Sciences, University of Reading, Reading, UK
Katie L. H. Gray
School of Psychology, University of Leeds, Leeds, LS2 9JT, UK
Richard Cook
You can also search for this author in PubMed Google Scholar
B.K.G., C.P., K.L.H.G., and R.C. designed the study. B.K.G. collected the data. B.K.G., K.L.H.G. and R.C. analysed the data. B.K.G., C.P., K.L.H.G., and R.C. wrote the manuscript.
Correspondence to Richard Cook .
Competing interests.
The authors declare no competing interests.
Publisher's note.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information., rights and permissions.
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ .
Reprints and permissions
Cite this article.
Gehdu, B.K., Press, C., Gray, K.L.H. et al. Autistic adults have insight into their relative face recognition ability. Sci Rep 14 , 17802 (2024). https://doi.org/10.1038/s41598-024-67649-8
Download citation
Received : 20 December 2023
Accepted : 15 July 2024
Published : 01 August 2024
DOI : https://doi.org/10.1038/s41598-024-67649-8
Anyone you share the following link with will be able to read this content:
Sorry, a shareable link is not currently available for this article.
Provided by the Springer Nature SharedIt content-sharing initiative
By submitting a comment you agree to abide by our Terms and Community Guidelines . If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.
Sign up for the Nature Briefing newsletter — what matters in science, free to your inbox daily.
New citation alert added.
This alert has been successfully added and will be sent to:
You will be notified whenever a record that you have chosen has been cited.
To manage your alert preferences, click on the button below.
Please log in to your account
Bibliometrics & citations, view options, recommendations, the privacy-personalization paradox in mhealth services acceptance of different age groups.
We examine the privacy-personalization paradox in mHealth service acceptance.Trust can mediate the effects of perceived personalization and privacy concerns on acceptance intention.For younger potential users, acceptance intention is largely driven by ...
Biometric technology BT is a component of information security and person identification. Individual acceptance and adoption of BT is fundamental to successful implementation of BT by organisations. There has been a fairly moderate but improving pace of ...
Engagement with e-portfolios has been shown to improve students' learning. However, what influences students to accept e-portfolios is a question that needs careful study. The purpose of this study is to investigate the influence of Self-Efficacy, ...
Published in.
Elsevier Science Publishers B. V.
Netherlands
Author tags.
Other metrics, bibliometrics, article metrics.
Login options.
Check if you have access through your login credentials or your institution to get full access on this article.
Share this publication link.
Copying failed.
Affiliations, export citations.
We are preparing your search results for download ...
We will inform you here when the file is ready.
Your file of search results citations is now ready.
Your search export query has expired. Please try again.
A not-for-profit organization, IEEE is the world's largest technical professional organization dedicated to advancing technology for the benefit of humanity. © Copyright 2024 IEEE - All rights reserved. Use of this web site signifies your agreement to the terms and conditions.
Face recognition device market to reach $16.5 billion, globally, by 2032 at 15.7% cagr: allied market research.
The global face recognition device market is experiencing growth driven by several factors, including the growing adoption in retail and e-commerce, expansion of smart city initiatives, and rising integration with IoT devices and other emerging technologies within the industry.
Wilmington, Delaware , Aug. 07, 2024 (GLOBE NEWSWIRE) -- Allied Market Research published a report, titled, " Face Recognition Device Market by Type (Standalone Devices and Integrated Devices), and Application (Surveillance, Access Control, Healthcare, Banking and Finance, Retail and E-commerce and Others): Global Opportunity Analysis and Industry Forecast, 2024-2032" . According to the report, the face recognition device market was valued at $4.5 billion in 2023, and is estimated to reach $16.5 billion by 2032, growing at a CAGR of 15.7% from 2024 to 2032.
Download Research Report Sample & TOC: https://www.alliedmarketresearch.com/request-sample/A68885
(We are providing report as per your research requirement, including the Latest Industry Insight's Evolution, Potential and COVID-19 Impact Analysis)
105 – Tables
57 – Charts
310 – Pages
Prime determinants of growth
The global face recognition device market is experiencing growth due to several factors, including the increasing adoption in retail and e-commerce and the expansion of smart city initiatives. However, accuracy issues somewhat hinder market growth. Moreover, the integration with IoT and other emerging technologies offers lucrative opportunities for the expansion of the global face recognition device market.
Report coverage & details:
|
|
Forecast Period | 2024–2032 |
Base Year | 2023 |
Market Size in 2023 | $4.5 billion |
Market Size in 2032 | $16.5 billion |
CAGR | 15.7% |
Segments Covered | Device Type, Application, and Region. |
Drivers | |
Opportunities | Integration with IoT and Other Emerging Technologies |
Restraint | Accuracy Issues |
Segment Highlights:
By device type, the global face recognition device market is bifurcated into standalone devices and integrated devices. Integrated devices lead the global face recognition device market. This dominance is driven by their seamless incorporation into a wide range of consumer electronics and security systems, benefiting from ongoing technological advancements in AI and machine learning.
By application, the global face recognition device market is divided into security and surveillance, access control, healthcare, banking and finance, retail and E-commerce, and others. The security and surveillance lead the global face recognition device market. This dominance is attributed to the increasing need for enhanced security measures across public and private sectors, including government buildings, airports, and critical infrastructure.
Get Customized Reports with your Requirements: https://www.alliedmarketresearch.com/request-for-customization/A68885
Region/Country Outlook:
The Asia-Pacific region leads the face recognition device market, with China at the forefront. This leadership is driven by substantial government investments in surveillance infrastructure, widespread adoption of facial recognition technology in public security and law enforcement, and a rapidly growing consumer electronics market.
Leading Market Players:
NEC Corporation
IDEMIA Group
Zhejiang Dahua Technology Co., Ltd.
Hangzhou Hikvision Digital Technology Co., Ltd.
Panasonic Corporation
Honeywell International Inc.
Fujitsu Limited.
Anviz Global Inc.
Sense Time Group Limited
VIVOTEK Inc.
The report provides a detailed analysis of these key players in the global face recognition device market. These players have adopted different strategies such as new product launches, collaborations, and others to increase their market share and maintain dominant shares in different regions. The report is valuable in highlighting business performance, operating segments, product portfolio, and strategic moves of market players to showcase the competitive scenario.
Key Industry Developments:
March 2024: Hikvision, a leading provider of security products, launched a new range of face recognition cameras designed for enhanced security in urban surveillance and critical infrastructure protection, boasting higher resolution and faster processing speeds.
November 2022: NEC Corporation developed a gateless access control system using biometric recognition that combines NEC's face recognition technology with person re-identification technology, which matches people even if they are facing away or their bodies are occluded, to provide fast and reliable entry control that is free from gates.
Inquiry before Buying: https://www.alliedmarketresearch.com/purchase-enquiry/A68885
Key Benefits For Stakeholders:
This report provides a quantitative analysis of the face recognition devices market segments, current trends, estimations, and dynamics of the market analysis to identify the prevailing market opportunities.
The market research is offered along with information related to key drivers, restraints, and opportunities.
Porter's five forces analysis highlights the potency of buyers and suppliers to enable stakeholders make profit-oriented business decisions and strengthen their supplier-buyer network.
In-depth analysis of the face recognition devices market segmentation assists to determine the prevailing market opportunities.
Major countries in each region are mapped according to their revenue contribution to the global face recognition devices market statistics.
Market player positioning facilitates benchmarking and provides a clear understanding of the present position of the market players.
The report includes the analysis of the regional as well as global f face recognition devices market trends, key players, market segments, application areas, and market growth strategies.
Procure Complete Report (250 Pages PDF with Insights, Charts, Tables, and Figures) @ https://www.alliedmarketresearch.com/checkout-final /face-recognition-device-market
Face Recognition Device Market Key Segments:
Standalone Devices
Integrated Devices
By Application
Surveillance
Access Control
Banking and Finance
Retail and E-commerce
North America (U.S., Canada, Mexico)
Europe (UK, Germany, France, Italy, Rest of Europe)
Asia- Pacific (China, Japan, India, South Korea, Rest of Asia-Pacific)
Latin America (Brazil, Argentina, Rest of Latin America)
Middle East and Africa (UAE, Saudi Arabia, Rest of Middle East And Africa)
Access AVENUE - A Subscription-Based Library (Premium On-Demand, Subscription-Based Pricing Model) @ https://www.alliedmarketresearch.com/library-access
Avenue is a user-based library of global market report database, provides comprehensive reports pertaining to the world's largest emerging markets. It further offers e-access to all the available industry reports just in a jiffy. By offering core business insights on the varied industries, economies, and end users worldwide, Avenue ensures that the registered members get an easy as well as single gateway to their all-inclusive requirements.
Avenue Library Subscription | Request For 14 Days Free Trial of Before Buying: https://www.alliedmarketresearch.com/avenue/trial/starter
Trending Reports in Semiconductor and Electronics Industry:
Industrial High Voltage Motor Market size was valued at $1.8 billion in 2022, and is projected to reach $2.6 billion by 2032, growing at a CAGR of 3.9% from 2023 to 2032.
High Voltage Capacitor Market was valued at $11.8 billion in 2020, and is projected to reach $30.3 billion by 2030, growing at a CAGR of 9.9% from 2021 to 2030.
Hearables Market size was valued at $21.20 billion in 2018, and is projected to reach $93.90 billion by 2026, growing at a CAGR of 17.2% from 2019 to 2026
Professional Portable Audio System Market was valued at $2.3 billion in 2021, and is projected to reach $5.1 billion by 2031, growing at a CAGR of 8.5% from 2022 to 2031
Allied Market Research (AMR) is a full-service market research and business-consulting wing of Allied Analytics LLP based in Wilmington, Delaware. Allied Market Research provides global enterprises as well as medium and small businesses with unmatched quality of "Market Research Reports Insights" and "Business Intelligence Solutions." AMR has a targeted view to provide business insights and consulting to assist its clients to make strategic business decisions and achieve sustainable growth in their respective market domain.
We are in professional corporate relations with various companies, and this helps us in digging out market data that helps us generate accurate research data tables and confirms utmost accuracy in our market forecasting. Allied Market Research CEO Pawan Kumar is instrumental in inspiring and encouraging everyone associated with the company to maintain high quality of data and help clients in every way possible to achieve success. Each and every data presented in the reports published by us is extracted through primary interviews with top officials from leading companies of domain concerned. Our secondary data procurement methodology includes deep online and offline research and discussion with knowledgeable professionals and analysts in the industry.
David Correa
1209 Orange Street, Corporation Trust Center, Wilmington, New Castle, Delaware 19801 USA. Int'l: +1-503-894-6022 Toll Free: +1-800-792-5285 UK: +44-845-528-1300 India (Pune): +91-20-66346060 Fax: +1-800-792-5285 [email protected]
By Paolo Vilbon, The Gavel, Contributor J.D. Candidate, Class of 2024
Artificial intelligence is the future and there is no denying that. But, with great advancements also comes the potential dangers associated with them. One of law enforcements biggest technological advancements in the last two decades has been the use of facial recognition technology. This coupled with modern artificial intelligence would lead some to think that this system would completely revolutionize law enforcement investigations and their standard operating procedures. Unfortunately, this is not the case. According to researchers, facial recognition technologies falsely identified Black and Asian faces 10 to 100 times more often than they did White faces. The technologies also falsely identified women more than they did men—making Black women particularly vulnerable to algorithmic bias.1
These algorithms currently help national agencies identify potential flight risks and protect borders.2 National agencies have an advantage over local law enforcement agencies because they possess the resources to cross check any information they receive, but local agencies do not have that kind of bandwidth. Further, it is no secret that efforts to recruit law enforcement officers have been on a downturn in recent years.3 This will lead police departments to rely more heavily on these technologies to fight crime. As the use of these systems increases, so will the errors associated with them. Therefore, if these technologies are not accurate or contain identifiable biases, they may do more harm than good.
One of the issues identified with artificial intelligence and facial detection is that AI face recognition tools “rely on machine learning algorithms that are trained with labeled data.”4 Further, “[i]t has recently been shown that algorithms trained with biased data have resulted in algorithmic discrimination.”5 The potential dangers associated with erroneous identification range from “missed flights, lengthy interrogations, watch list placements, tense police encounters, false arrests, or worse.”6 All of which ignore the financial impact that a false identification will have on the individual. Society must hold companies who put face recognition tools into the marketplace accountable in the hopes that new development of technologies will be much more accurate. This would ensure that future algorithms will prevent harm to the individuals that these technologies are biased against.
Artificial intelligence is far too embedded into daily life to slow its progress but claims that the data set used for its baselines is biased should not be ignored. These biases should be brought to the forefront so that the necessary changes can be made now before artificial intelligence needlessly overburdens the criminal justice system. A yearlong research investigation across 100 police departments revealed that African American individuals are more likely to be stopped by law enforcement and be subjected to face recognition searches than individuals of other ethnicities.7 This happens because, without a dataset that has labels for various skin characteristics such as color, thickness, and the amount of hair, one cannot measure the accuracy of such automated detection systems. Although it may sound ridiculous, we are at a turning point when it comes to this technology. If this technology is continuously used with the current biases it has, it will be useful, but will also lead to mass incarceration of the wrong suspects. This will then negatively harm the government and impacted individuals economically while also carrying a negative social impact. It is imperative that we realize that these biases exist so they can be corrected now.
References:
1 The Regulatory Review. “Saturday Seminar: Facing Bias in Facial Recognition Technology.”
3 U.S. Experiencing Police Hiring Crisis
4 Gender Shades: Intersectional Accuracy Disparities in Commercial Gender Classification.
Does artificially manufactured art satisfy the originality element in copyright law, reshaping and enhancing the e-discovery process, a constitutional right, artificial intelligence and the legal world – will more class actions result, revolutionizing traditional methods of legal research.
Comments are closed.
Fully accredited by the American Bar Association, Council of the Section of Legal Education and Admissions to the Bar of the American Bar Association, 321 N. Clark Street, Chicago, IL 60654, (312) 988-6738. Licensed by the Commission for Independent Education, Florida Department of Education, license # 4007. Additional information regarding this institution may be obtained by contacting the Commission at 325 West Gaines Street, Suite 1414, Tallahassee, FL, 32399-0400. Toll-free telephone number (888) 224-6684.
1025 Commons Cir, Naples, FL 34119
© 2024 Ave Maria School of Law. All Rights Reserved
(239) 687-5300
IMAGES
COMMENTS
Face recognition is one of the most active research fields of computer vision and pattern recognition, with many practical and commercial applications including identification, access control, forensics, and human-computer interactions. However, identifying a face in a crowd raises serious questions about individual freedoms and poses ethical issues. Significant methods, algorithms, approaches ...
A novel taxonomy of image and video-based methods, which also contains recent methods such as sparsity and deep learning based methods. An up-to-date review of the image and video-based data sets used for face recognition. Review of the recent deep-learning based methods, which have shown remarkable results on large scale and unconstrained ...
Abstract. Deep learning models currently achieve human levels of performance on real-world face recognition tasks. We review scientific progress in understanding human face processing using computational approaches based on deep learning. This review is organized around three fundamental advances.
Face recognition is an efficient technique and one of the most preferred biometric modalities for the identification and verification of individuals as compared to voice, fingerprint, iris, retina eye scan, gait, ear and hand geometry. This has over the years necessitated researchers in both the academia and industry to come up with several face recognition techniques making it one of the most ...
MIT neuroscientists have found that when artificial intelligence is tasked with visually identifying objects and faces, it assigns specific components of its network to face recognition - just like the human brain.
In this review, we have highlighted major applications, challenges and trends of face recognition systems in social and scientific domains. The prime objective of this research is to sum-up recent face recognition techniques and develop a broad understanding of how these techniques behave on different datasets.
Deep learning models currently achieve human levels of performance on real-world face recognition tasks. We review scientific progress in understanding human face processing using computational approaches based on deep learning. This review is organized around three fundamental advances. First, deep networks trained for face identification generate a representation that retains structured ...
Abstract Face recognition is an efficient technique and one of the most preferred biometric modalities for the identification and verification of individuals as compared to voice, fingerprint, iris, retina eye scan, gait, ear and hand geometry. This has over the years necessitated researchers in both the academia and industry to come up with several face recognition techniques making it one of ...
Abstract and Figures. Face recognition is an efficient technique and one of the most preferred biometric modalities for the identification and verification of individuals as compared to voice ...
Recent years witnessed the breakthrough of face recognition with deep convolutional neural networks. Dozens of papers in the field of FR are published every year. Some of them were applied in the industrial community and played an important role in human life such as device unlock, mobile payment, and so on. This paper provides an introduction to face recognition, including its history ...
A Review of Face Recognition Technology: In the previous few decades, face recognition has become a popular field in computer-based application development This is due to the fact that it is employed in so many different sectors. Face identification via database photographs, real data, captured images, and sensor images is also a difficult task due to the huge variety of faces. The fields of ...
The recognition rate and speed of face recognition systems have always been the two key technical factors that researchers focus on.
Description with markdown (optional): **Facial Recognition** is the task of making a positive identification of a face in a photo or video image against a pre-existing database of faces. It begins with detection - distinguishing human faces from other objects in the image - and then works on identification of those detected faces.
The aim of this focused review is to recount a string of key discoveries about individual differences in face recognition made during the last decade. Fig. 1. The recent acceleration of research on individual differences in face recognition. Each line represents averaged results for similar searches (producing similar results) in Google Scholar ...
ABSTRACT Face recognition is a rapidly developing and widely applied aspect of biometric technologies. Its applications are broad, ranging from law enforcement to consumer applications, and industry efficiency and monitoring solutions. The recent advent of affordable, powerful GPUs and the creation of huge face databases has drawn research focus primarily on the development of increasingly ...
Metrics. Abstract: Face recognition technology is a biometric technology, which is based on the identification of facial features of a person. People collect the face images, and the recognition equipment automatically processes the images. The paper introduces the related researches of face recognition from different perspectives.
To deal with the issue of human face recognition on small original dataset, a new approach combining convolutional neural network (CNN) with augmented dataset is developed in this paper. The origin...
These varied accomplishments across many distinct subdomains of face-recognition research are extensive and conveying the current state of knowledge and looking ahead to future challenges is difficult when there is this much material to draw upon.
This paper highlights the recent research on the 2D or 3D face recognition system, focusing mainly on approaches based on local, holistic (subspace), and hybrid features.
This paper highlights the recent research on the 2D or 3D face recognition system, focusing mainly on approaches based on local, holistic (subspace), and hybrid features.
Face recognition has long been an active research area in the field of artificial intelligence, particularly since the rise of deep learning in recent years. In some practical situations, each identity has only a single sample available for training. Face recognition under this situation is referred to as single sample face recognition and poses significant challenges to the effective training ...
Abstract and Figures The task of face recognition has been actively researched in recent years. This paper provides an up-to-date review of major human face recognition research.
This paper investigates various feature-based automatic face recognition approaches in detail. High degree of freedom in head movement and human emotion leads a face recognition system to face critical challenges in terms of pose, illumination and expression. Human face also undergoes irreversible changes due to aging.
Recent reports suggest that the PI20 scores of autistic participants exhibit little or no correlation with their performance on the Cambridge Face Memory Test—a key measure of face recognition ...
In this study, we investigated the shared neural dynamics of emotional face processing using an explicit facial emotion recognition task, where participants made two-alternative forced choice (2AFC) decisions on the displayed emotion.
In this paper, we present a method of expression-invariant face recognition that transforms input face image with an arbitrary expression into its corresponding neutral facial expression image. When a new face image with an arbitrary expression is ...
C. Morosan, Hotel facial recognition systems: Insight into guests' system perceptions, congruity with self-image, and anticipated emotions, Journal of Electronic Commerce Research 21 (1) (2020) 21-38.
This Facial emotion recognition plays a pivotal role in human-computer interaction, with far-reaching implications spanning psychology, healthcare, and human computer interaction. Conventional approaches, primarily utilizing classical machine learning methods, frequently face challenges in effectively interpreting intricate facial expressions under diverse environmental circumstances. However ...
The global face recognition device market is experiencing growth driven by several factors, including the growing adoption in retail and e-commerce, expansion of smart city initiatives, and rising ...
A yearlong research investigation across 100 police departments revealed that African American individuals are more likely to be stopped by law enforcement and be subjected to face recognition searches than individuals of other ethnicities.7 This happens because, without a dataset that has labels for various skin characteristics such as color ...