Sparse representation is an effective data classification algorithm that depends on the known training samples to categorise the test sample.It has been widely used in various image classification tasks.Sparseness in ...Sparse representation is an effective data classification algorithm that depends on the known training samples to categorise the test sample.It has been widely used in various image classification tasks.Sparseness in sparse representation means that only a few of instances selected from all training samples can effectively convey the essential class-specific information of the test sample,which is very important for classification.For deformable images such as human faces,pixels at the same location of different images of the same subject usually have different intensities.Therefore,extracting features and correctly classifying such deformable objects is very hard.Moreover,the lighting,attitude and occlusion cause more difficulty.Considering the problems and challenges listed above,a novel image representation and classification algorithm is proposed.First,the authors’algorithm generates virtual samples by a non-linear variation method.This method can effectively extract the low-frequency information of space-domain features of the original image,which is very useful for representing deformable objects.The combination of the original and virtual samples is more beneficial to improve the clas-sification performance and robustness of the algorithm.Thereby,the authors’algorithm calculates the expression coefficients of the original and virtual samples separately using the sparse representation principle and obtains the final score by a designed efficient score fusion scheme.The weighting coefficients in the score fusion scheme are set entirely automatically.Finally,the algorithm classifies the samples based on the final scores.The experimental results show that our method performs better classification than conventional sparse representation algorithms.展开更多
Corona virus(COVID-19)is once in a life time calamity that has resulted in thousands of deaths and security concerns.People are using face masks on a regular basis to protect themselves and to help reduce corona virus...Corona virus(COVID-19)is once in a life time calamity that has resulted in thousands of deaths and security concerns.People are using face masks on a regular basis to protect themselves and to help reduce corona virus transmission.During the on-going coronavirus outbreak,one of the major priorities for researchers is to discover effective solution.As important parts of the face are obscured,face identification and verification becomes exceedingly difficult.The suggested method is a transfer learning using MobileNet V2 based technology that uses deep feature such as feature extraction and deep learning model,to identify the problem of face masked identification.In the first stage,we are applying face mask detector to identify the face mask.Then,the proposed approach is applying to the datasets from Canadian Institute for Advanced Research10(CIFAR10),Modified National Institute of Standards and Technology Database(MNIST),Real World Masked Face Recognition Database(RMFRD),and Stimulated Masked Face Recognition Database(SMFRD).The proposed model is achieving recognition accuracy 99.82%with proposed dataset.This article employs the four pre-programmed models VGG16,VGG19,ResNet50 and ResNet101.To extract the deep features of faces with VGG16 is achieving 99.30%accuracy,VGG19 is achieving 99.54%accuracy,ResNet50 is achieving 78.70%accuracy and ResNet101 is achieving 98.64%accuracy with own dataset.The comparative analysis shows,that our proposed model performs better result in all four previous existing models.The fundamental contribution of this study is to monitor with face mask and without face mask to decreases the pace of corona virus and to detect persons using wearing face masks.展开更多
Convolutional neural networks continually evolve to enhance accuracy in addressing various problems,leading to an increase in computational cost and model size.This paper introduces a novel approach for pruning face r...Convolutional neural networks continually evolve to enhance accuracy in addressing various problems,leading to an increase in computational cost and model size.This paper introduces a novel approach for pruning face recognition models based on convolutional neural networks.The proposed method identifies and removes inefficient filters based on the information volume in feature maps.In each layer,some feature maps lack useful information,and there exists a correlation between certain feature maps.Filters associated with these two types of feature maps impose additional computational costs on the model.By eliminating filters related to these categories of feature maps,the reduction of both computational cost and model size can be achieved.The approach employs a combination of correlation analysis and the summation of matrix elements within each feature map to detect and eliminate inefficient filters.The method was applied to two face recognition models utilizing the VGG16 and ResNet50V2 backbone architectures.In the proposed approach,the number of filters removed in each layer varies,and the removal process is independent of the adjacent layers.The convolutional layers of both backbone models were initialized with pre-trained weights from ImageNet.For training,the CASIA-WebFace dataset was utilized,and the Labeled Faces in the Wild(LFW)dataset was employed for benchmarking purposes.In the VGG16-based face recognition model,a 0.74%accuracy improvement was achieved while reducing the number of convolution parameters by 26.85%and decreasing Floating-point operations per second(FLOPs)by 47.96%.For the face recognition model based on the ResNet50V2 architecture,the ArcFace method was implemented.The removal of inactive filters in this model led to a slight decrease in accuracy by 0.11%.However,it resulted in enhanced training speed,a reduction of 59.38%in convolution parameters,and a 57.29%decrease in FLOPs.展开更多
Support vector machine (SVM), as a novel approach in pattern recognition, has demonstrated a success in face detection and face recognition. In this paper, a face recognition approach based on the SVM classifier with ...Support vector machine (SVM), as a novel approach in pattern recognition, has demonstrated a success in face detection and face recognition. In this paper, a face recognition approach based on the SVM classifier with the nearest neighbor classifier (NNC) is proposed. The principal component analysis (PCA) is used to reduce the dimension and extract features. Then one-against-all stratedy is used to train the SVM classifiers. At the testing stage, we propose an al-展开更多
Dimensionality reduction methods play an important role in face recognition. Principal component analysis(PCA) and two-dimensional principal component analysis(2DPCA) are two kinds of important methods in this field. ...Dimensionality reduction methods play an important role in face recognition. Principal component analysis(PCA) and two-dimensional principal component analysis(2DPCA) are two kinds of important methods in this field. Recent research seems like that 2DPCA method is superior to PCA method. To prove if this conclusion is always true, a comprehensive comparison study between PCA and 2DPCA methods was carried out. A novel concept, called column-image difference(CID), was proposed to analyze the difference between PCA and 2DPCA methods in theory. It is found that there exist some restrictive conditions when2 DPCA outperforms PCA. After theoretical analysis, the experiments were conducted on four famous face image databases. The experiment results confirm the validity of theoretical claim.展开更多
Face recognition provides a natural visual interface for human computer interaction (HCI) applications. The process of face recognition, however, is inhibited by variations in the appearance of face images caused by...Face recognition provides a natural visual interface for human computer interaction (HCI) applications. The process of face recognition, however, is inhibited by variations in the appearance of face images caused by changes in lighting, expression, viewpoint, aging and introduction of occlusion. Although various algorithms have been presented for face recognition, face recognition is still a very challenging topic. A novel approach of real time face recognition for HCI is proposed in the paper. In view of the limits of the popular approaches to foreground segmentation, wavelet multi-scale transform based background subtraction is developed to extract foreground objects. The optimal selection of the threshold is automatically determined, which does not require any complex supervised training or manual experimental calibration. A robust real time face recognition algorithm is presented, which combines the projection matrixes without iteration and kernel Fisher discriminant analysis (KFDA) to overcome some difficulties existing in the real face recognition. Superior performance of the proposed algorithm is demonstrated by comparing with other algorithms through experiments. The proposed algorithm can also be applied to the video image sequences of natural HCI.展开更多
The purpose of this paper is to solve the problem of robust face recognition(FR) with single sample per person(SSPP). In the scenario of FR with SSPP, we present a novel model local robust sparse representation(LRSR) ...The purpose of this paper is to solve the problem of robust face recognition(FR) with single sample per person(SSPP). In the scenario of FR with SSPP, we present a novel model local robust sparse representation(LRSR) to tackle the problem of query images with various intra-class variations,e.g., expressions, illuminations, and occlusion. FR with SSPP is a very difficult challenge due to lacking of information to predict the possible intra-class variation of the query images.The key idea of the proposed method is to combine a local sparse representation model and a patch-based generic variation dictionary learning model to predict the possible facial intraclass variation of the query images. The experimental results on the AR database, Extended Yale B database, CMU-PIE database and LFW database show that the proposed method is robust to intra-class variations in FR with SSPP, and outperforms the state-of-art approaches.展开更多
Most face recognition techniques have been successful in dealing with high-resolution(HR) frontal face images. However, real-world face recognition systems are often confronted with the low-resolution(LR) face images ...Most face recognition techniques have been successful in dealing with high-resolution(HR) frontal face images. However, real-world face recognition systems are often confronted with the low-resolution(LR) face images with pose and illumination variations. This is a very challenging issue, especially under the constraint of using only a single gallery image per person.To address the problem, we propose a novel approach called coupled kernel-based enhanced discriminant analysis(CKEDA).CKEDA aims to simultaneously project the features from LR non-frontal probe images and HR frontal gallery ones into a common space where discrimination property is maximized.There are four advantages of the proposed approach: 1) by using the appropriate kernel function, the data becomes linearly separable, which is beneficial for recognition; 2) inspired by linear discriminant analysis(LDA), we integrate multiple discriminant factors into our objective function to enhance the discrimination property; 3) we use the gallery extended trick to improve the recognition performance for a single gallery image per person problem; 4) our approach can address the problem of matching LR non-frontal probe images with HR frontal gallery images,which is difficult for most existing face recognition techniques.Experimental evaluation on the multi-PIE dataset signifies highly competitive performance of our algorithm.展开更多
It has been demonstrated that the linear discriminant analysis (LDA) is an effective approach in face recognition tasks. However, due to the high dimensionality of an image space, many LDA based approaches first use t...It has been demonstrated that the linear discriminant analysis (LDA) is an effective approach in face recognition tasks. However, due to the high dimensionality of an image space, many LDA based approaches first use the principal component analysis (PCA) to project an image into a lower dimensional space, then perform the LDA transform to extract discriminant feature. But some useful discriminant information to the following LDA transform will be lost in the PCA step. To overcome these defects, a face recognition method based on the discrete cosine transform (DCT) and the LDA is proposed. First the DCT is used to achieve dimension reduction, then LDA transform is performed on the lower space to extract features. Two face databases are used to test our method and the correct recognition rates of 97.5% and 96.0% are obtained respectively. The performance of the proposed method is compared with that of the PCA+ LDA method and the results show that the method proposed outperforms the PCA+ LDA method.展开更多
In order to distinguish faces of various angles during face recognition, an algorithm of the combination of approximate dynamic programming (ADP) called action dependent heuristic dynamic programming (ADHDP) and p...In order to distinguish faces of various angles during face recognition, an algorithm of the combination of approximate dynamic programming (ADP) called action dependent heuristic dynamic programming (ADHDP) and particle swarm optimization (PSO) is presented. ADP is used for dynamically changing the values of the PSO parameters. During the process of face recognition, the discrete cosine transformation (DCT) is first introduced to reduce negative effects. Then, Karhunen-Loeve (K-L) transformation can be used to compress images and decrease data dimensions. According to principal component analysis (PCA), the main parts of vectors are extracted for data representation. Finally, radial basis function (RBF) neural network is trained to recognize various faces. The training of RBF neural network is exploited by ADP-PSO. In terms of ORL Face Database, the experimental result gives a clear view of its accurate efficiency.展开更多
Face recognition based on few training samples is a challenging task. In daily applications, sufficient training samples may not be obtained and most of the gained training samples are in various illuminations and pos...Face recognition based on few training samples is a challenging task. In daily applications, sufficient training samples may not be obtained and most of the gained training samples are in various illuminations and poses. Non-sufficient training samples could not effectively express various facial conditions, so the improvement of the face recognition rate under the non-sufficient training samples condition becomes a laborious mission. In our work, the facial pose pre-recognition(FPPR) model and the dualdictionary sparse representation classification(DD-SRC) are proposed for face recognition. The FPPR model is based on the facial geometric characteristic and machine learning, dividing a testing sample into full-face and profile. Different poses in a single dictionary are influenced by each other, which leads to a low face recognition rate. The DD-SRC contains two dictionaries, full-face dictionary and profile dictionary, and is able to reduce the interference. After FPPR, the sample is processed by the DD-SRC to find the most similar one in training samples. The experimental results show the performance of the proposed algorithm on olivetti research laboratory(ORL) and face recognition technology(FERET) databases, and also reflect comparisons with SRC, linear regression classification(LRC), and two-phase test sample sparse representation(TPTSSR).展开更多
An improved face recognition method is proposed based on principal component analysis (PCA) compounded with genetic algorithm (GA), named as genetic based principal component analysis (GPCA). Initially the eigen...An improved face recognition method is proposed based on principal component analysis (PCA) compounded with genetic algorithm (GA), named as genetic based principal component analysis (GPCA). Initially the eigenspace is created with eigenvalues and eigenvectors. From this space, the eigenfaces are constructed, and the most relevant eigenfaees have been selected using GPCA. With these eigenfaees, the input images are classified based on Euclidian distance. The proposed method was tested on ORL (Olivetti Research Labs) face database. Experimental results on this database demonstrate that the effectiveness of the proposed method for face recognition has less misclassification in comparison with previous methods.展开更多
Since the coal mine in-pit personnel positioning system neither can effectively achieve the function to detect the uniqueness of in-pit coal-mine personnel nor can identify and eliminate violations in attendance manag...Since the coal mine in-pit personnel positioning system neither can effectively achieve the function to detect the uniqueness of in-pit coal-mine personnel nor can identify and eliminate violations in attendance management such as multiple cards for one person, and swiping one's cards by others in China at present. Therefore, the research introduces a uniqueness detection system and method for in-pit coal-mine personnel integrated into the in-pit coal mine personnel positioning system, establishing a system mode based on face recognition + recognition of personnel positioning card + release by automatic detection. Aiming at the facts that the in-pit personnel are wearing helmets and faces are prone to be stained during the face recognition, the study proposes the ideas that pre-process face images using the 2D-wavelet-transformation-based Mallat algorithm and extracts three face features: miner light, eyes and mouths, using the generalized symmetry transformation-based algorithm. This research carried out test with 40 clean face images with no helmets and 40 lightly-stained face images, and then compared with results with the one using the face feature extraction method based on grey-scale transformation and edge detection. The results show that the method described in the paper can detect accurately face features in the above-mentioned two cases, and the accuracy to detect face features is 97.5% in the case of wearing helmets and lightly-stained faces.展开更多
Face recognition systems have enhanced human-computer interactions in the last ten years.However,the literature reveals that current techniques used for identifying or verifying faces are not immune to limitations.Pri...Face recognition systems have enhanced human-computer interactions in the last ten years.However,the literature reveals that current techniques used for identifying or verifying faces are not immune to limitations.Principal Component Analysis-Support Vector Machine(PCA-SVM)and Principal Component Analysis-Artificial Neural Network(PCA-ANN)are among the relatively recent and powerful face analysis techniques.Compared to PCA-ANN,PCA-SVM has demonstrated generalization capabilities in many tasks,including the ability to recognize objects with small or large data samples.Apart from requiring a minimal number of parameters in face detection,PCA-SVM minimizes generalization errors and avoids overfitting problems better than PCA-ANN.PCA-SVM,however,is ineffective and inefficient in detecting human faces in cases in which there is poor lighting,long hair,or items covering the subject’s face.This study proposes a novel PCASVM-based model to overcome the recognition problem of PCA-ANN and enhance face detection.The experimental results indicate that the proposed model provides a better face recognition outcome than PCA-SVM.展开更多
An algorithm for face description and recognition based on multi-resolution with multi-scale local binary pattern (multi-LBP) features is proposed. The facial image pyramid is constructed and each facial image is di...An algorithm for face description and recognition based on multi-resolution with multi-scale local binary pattern (multi-LBP) features is proposed. The facial image pyramid is constructed and each facial image is divided into various regions from which partial and holistic local binary patter (LBP) histograms are extracted. All LBP features of each image are concatenated to a single LBP eigenvector with different resolutions. The dimensionaUty of LBP features is then reduced by a local margin alignment (LMA) algorithm based on manifold, which can preserve the between-class variance. Support vector machine (SVM) is applied to classify facial images. Extensive experiments on ORL and CMU face databases clearly show the superiority of the proposed scheme over some existed algorithms, especially on the robustness of the method against different facial expressions and postures of the subjects.展开更多
Over the past few decades,face recognition has become the most effective biometric technique in recognizing people’s identity,as it is widely used in many areas of our daily lives.However,it is a challenging techniqu...Over the past few decades,face recognition has become the most effective biometric technique in recognizing people’s identity,as it is widely used in many areas of our daily lives.However,it is a challenging technique since facial images vary in rotations,expressions,and illuminations.To minimize the impact of these challenges,exploiting information from various feature extraction methods is recommended since one of the most critical tasks in face recognition system is the extraction of facial features.Therefore,this paper presents a new approach to face recognition based on the fusion of Gabor-based feature extraction,Fast Independent Component Analysis(FastICA),and Linear Discriminant Analysis(LDA).In the presented method,first,face images are transformed to grayscale and resized to have a uniform size.After that,facial features are extracted from the aligned face image using Gabor,FastICA,and LDA methods.Finally,the nearest distance classifier is utilized to recognize the identity of the individuals.Here,the performance of six distance classifiers,namely Euclidean,Cosine,Bray-Curtis,Mahalanobis,Correlation,and Manhattan,are investigated.Experimental results revealed that the presented method attains a higher rank-one recognition rate compared to the recent approaches in the literature on four benchmarked face datasets:ORL,GT,FEI,and Yale.Moreover,it showed that the proposed method not only helps in better extracting the features but also in improving the overall efficiency of the facial recognition system.展开更多
In order to solve the problem that real-time face recognition is susceptible to illumination changes,this paper proposes a face recognition method that combines Local Binary Patterns(LBP)and Embedded Hidden Markov Mod...In order to solve the problem that real-time face recognition is susceptible to illumination changes,this paper proposes a face recognition method that combines Local Binary Patterns(LBP)and Embedded Hidden Markov Model(EHMM).Face recognition method.The method firstly performs LBP preprocessing on the input face image,then extracts the feature vector,and finally sends the extracted feature observation vector to the EHMM for training or recognition.Experiments on multiple face databases show that the proposed algorithm is robust to illumination and improves recognition rate.展开更多
This research presents an improved real-time face recognition system at a low<span><span><span style="font-family:" color:red;"=""> </span></span></span><...This research presents an improved real-time face recognition system at a low<span><span><span style="font-family:" color:red;"=""> </span></span></span><span style="font-family:Verdana;"><span style="font-family:Verdana;"><span style="font-family:Verdana;">resolution of 15 pixels with pose and emotion and resolution variations. We have designed our datasets named LRD200 and LRD100, which have been used for training and classification. The face detection part uses the Viola-Jones algorithm, and the face recognition part receives the face image from the face detection part to process it using the Local Binary Pattern Histogram (LBPH) algorithm with preprocessing using contrast limited adaptive histogram equalization (CLAHE) and face alignment. The face database in this system can be updated via our custom-built standalone android app and automatic restarting of the training and recognition process with an updated database. Using our proposed algorithm, a real-time face recognition accuracy of 78.40% at 15</span></span></span><span><span><span style="font-family:;" "=""> </span></span></span><span style="font-family:Verdana;"><span style="font-family:Verdana;"><span style="font-family:Verdana;">px and 98.05% at 45</span></span></span><span><span><span style="font-family:;" "=""> </span></span></span><span style="font-family:Verdana;"><span style="font-family:Verdana;"><span style="font-family:Verdana;">px have been achieved using the LRD200 database containing 200 images per person. With 100 images per person in the database (LRD100) the achieved accuracies are 60.60% at 15</span></span></span><span><span><span style="font-family:;" "=""> </span></span></span><span style="font-family:Verdana;"><span style="font-family:Verdana;"><span style="font-family:Verdana;">px and 95% at 45</span></span></span><span><span><span style="font-family:;" "=""> </span></span></span><span style="font-family:Verdana;"><span style="font-family:Verdana;"><span style="font-family:Verdana;">px respectively. A facial deflection of about 30</span></span></span><span><span><span><span><span style="color:#4F4F4F;font-family:-apple-system, " font-size:16px;white-space:normal;background-color:#ffffff;"="">°</span></span><span> on either side from the front face showed an average face recognition precision of 72.25%-81.85%. This face recognition system can be employed for law enforcement purposes, where the surveillance camera captures a low-resolution image because of the distance of a person from the camera. It can also be used as a surveillance system in airports, bus stations, etc., to reduce the risk of possible criminal threats.</span></span></span></span>展开更多
Time lapse, characteristic of aging, is a complex process that affects the reliability and security of biometric face recognition systems. This paper reports the novel use and effectiveness of deep learning, in genera...Time lapse, characteristic of aging, is a complex process that affects the reliability and security of biometric face recognition systems. This paper reports the novel use and effectiveness of deep learning, in general, and convolutional neural networks (CNN), in particular, for automatic rather than hand-crafted feature extraction for robust face recognition across time lapse. A CNN architecture using the VGG-Face deep (neural network) learning is found to produce highly discriminative and interoperable features that are robust to aging variations even across a mix of biometric datasets. The features extracted show high inter-class and low intra-class variability leading to low generalization errors on aging datasets using ensembles of subspace discriminant classifiers. The classification results for the all-encompassing authentication methods proposed on the challenging FG-NET and MORPH datasets are competitive with state-of-the-art methods including commercial face recognition engines and are richer in functionality and interoperability than existing methods as it handles mixed biometric datasets, e.g., FG-NET and MORPH.展开更多
A face recognition scheme is proposed, wherein a face image is preprocessed by pixel averaging and energy normalizing to reduce data dimension and brightness variation effect, followed by the Fourier transform to esti...A face recognition scheme is proposed, wherein a face image is preprocessed by pixel averaging and energy normalizing to reduce data dimension and brightness variation effect, followed by the Fourier transform to estimate the spectrum of the preprocessed image. The principal component analysis is conducted on the spectra of a face image to obtain eigen features. Combining eigen features with a Parzen classifier, experiments are taken on the ORL face database.展开更多
文摘Sparse representation is an effective data classification algorithm that depends on the known training samples to categorise the test sample.It has been widely used in various image classification tasks.Sparseness in sparse representation means that only a few of instances selected from all training samples can effectively convey the essential class-specific information of the test sample,which is very important for classification.For deformable images such as human faces,pixels at the same location of different images of the same subject usually have different intensities.Therefore,extracting features and correctly classifying such deformable objects is very hard.Moreover,the lighting,attitude and occlusion cause more difficulty.Considering the problems and challenges listed above,a novel image representation and classification algorithm is proposed.First,the authors’algorithm generates virtual samples by a non-linear variation method.This method can effectively extract the low-frequency information of space-domain features of the original image,which is very useful for representing deformable objects.The combination of the original and virtual samples is more beneficial to improve the clas-sification performance and robustness of the algorithm.Thereby,the authors’algorithm calculates the expression coefficients of the original and virtual samples separately using the sparse representation principle and obtains the final score by a designed efficient score fusion scheme.The weighting coefficients in the score fusion scheme are set entirely automatically.Finally,the algorithm classifies the samples based on the final scores.The experimental results show that our method performs better classification than conventional sparse representation algorithms.
文摘Corona virus(COVID-19)is once in a life time calamity that has resulted in thousands of deaths and security concerns.People are using face masks on a regular basis to protect themselves and to help reduce corona virus transmission.During the on-going coronavirus outbreak,one of the major priorities for researchers is to discover effective solution.As important parts of the face are obscured,face identification and verification becomes exceedingly difficult.The suggested method is a transfer learning using MobileNet V2 based technology that uses deep feature such as feature extraction and deep learning model,to identify the problem of face masked identification.In the first stage,we are applying face mask detector to identify the face mask.Then,the proposed approach is applying to the datasets from Canadian Institute for Advanced Research10(CIFAR10),Modified National Institute of Standards and Technology Database(MNIST),Real World Masked Face Recognition Database(RMFRD),and Stimulated Masked Face Recognition Database(SMFRD).The proposed model is achieving recognition accuracy 99.82%with proposed dataset.This article employs the four pre-programmed models VGG16,VGG19,ResNet50 and ResNet101.To extract the deep features of faces with VGG16 is achieving 99.30%accuracy,VGG19 is achieving 99.54%accuracy,ResNet50 is achieving 78.70%accuracy and ResNet101 is achieving 98.64%accuracy with own dataset.The comparative analysis shows,that our proposed model performs better result in all four previous existing models.The fundamental contribution of this study is to monitor with face mask and without face mask to decreases the pace of corona virus and to detect persons using wearing face masks.
文摘Convolutional neural networks continually evolve to enhance accuracy in addressing various problems,leading to an increase in computational cost and model size.This paper introduces a novel approach for pruning face recognition models based on convolutional neural networks.The proposed method identifies and removes inefficient filters based on the information volume in feature maps.In each layer,some feature maps lack useful information,and there exists a correlation between certain feature maps.Filters associated with these two types of feature maps impose additional computational costs on the model.By eliminating filters related to these categories of feature maps,the reduction of both computational cost and model size can be achieved.The approach employs a combination of correlation analysis and the summation of matrix elements within each feature map to detect and eliminate inefficient filters.The method was applied to two face recognition models utilizing the VGG16 and ResNet50V2 backbone architectures.In the proposed approach,the number of filters removed in each layer varies,and the removal process is independent of the adjacent layers.The convolutional layers of both backbone models were initialized with pre-trained weights from ImageNet.For training,the CASIA-WebFace dataset was utilized,and the Labeled Faces in the Wild(LFW)dataset was employed for benchmarking purposes.In the VGG16-based face recognition model,a 0.74%accuracy improvement was achieved while reducing the number of convolution parameters by 26.85%and decreasing Floating-point operations per second(FLOPs)by 47.96%.For the face recognition model based on the ResNet50V2 architecture,the ArcFace method was implemented.The removal of inactive filters in this model led to a slight decrease in accuracy by 0.11%.However,it resulted in enhanced training speed,a reduction of 59.38%in convolution parameters,and a 57.29%decrease in FLOPs.
基金This project was supported by Shanghai Shu Guang Project.
文摘Support vector machine (SVM), as a novel approach in pattern recognition, has demonstrated a success in face detection and face recognition. In this paper, a face recognition approach based on the SVM classifier with the nearest neighbor classifier (NNC) is proposed. The principal component analysis (PCA) is used to reduce the dimension and extract features. Then one-against-all stratedy is used to train the SVM classifiers. At the testing stage, we propose an al-
基金Projects(50275150,61173052)supported by the National Natural Science Foundation of China
文摘Dimensionality reduction methods play an important role in face recognition. Principal component analysis(PCA) and two-dimensional principal component analysis(2DPCA) are two kinds of important methods in this field. Recent research seems like that 2DPCA method is superior to PCA method. To prove if this conclusion is always true, a comprehensive comparison study between PCA and 2DPCA methods was carried out. A novel concept, called column-image difference(CID), was proposed to analyze the difference between PCA and 2DPCA methods in theory. It is found that there exist some restrictive conditions when2 DPCA outperforms PCA. After theoretical analysis, the experiments were conducted on four famous face image databases. The experiment results confirm the validity of theoretical claim.
基金supported by the National Natural Science Foundation of China (Grant No.60872117)the Leading Academic Discipline Project of Shanghai Municipal Education Commission (Grant No.J50104)
文摘Face recognition provides a natural visual interface for human computer interaction (HCI) applications. The process of face recognition, however, is inhibited by variations in the appearance of face images caused by changes in lighting, expression, viewpoint, aging and introduction of occlusion. Although various algorithms have been presented for face recognition, face recognition is still a very challenging topic. A novel approach of real time face recognition for HCI is proposed in the paper. In view of the limits of the popular approaches to foreground segmentation, wavelet multi-scale transform based background subtraction is developed to extract foreground objects. The optimal selection of the threshold is automatically determined, which does not require any complex supervised training or manual experimental calibration. A robust real time face recognition algorithm is presented, which combines the projection matrixes without iteration and kernel Fisher discriminant analysis (KFDA) to overcome some difficulties existing in the real face recognition. Superior performance of the proposed algorithm is demonstrated by comparing with other algorithms through experiments. The proposed algorithm can also be applied to the video image sequences of natural HCI.
基金supported in part by the National Natural Science Foundation of China(61673402,61273270,60802069)the Natural Science Foundation of Guangdong Province(2017A030311029,2016B010109002,2015B090912001,2016B010123005,2017B090909005)+1 种基金the Science and Technology Program of Guangzhou of China(201704020180,201604020024)the Fundamental Research Funds for the Central Universities of China
文摘The purpose of this paper is to solve the problem of robust face recognition(FR) with single sample per person(SSPP). In the scenario of FR with SSPP, we present a novel model local robust sparse representation(LRSR) to tackle the problem of query images with various intra-class variations,e.g., expressions, illuminations, and occlusion. FR with SSPP is a very difficult challenge due to lacking of information to predict the possible intra-class variation of the query images.The key idea of the proposed method is to combine a local sparse representation model and a patch-based generic variation dictionary learning model to predict the possible facial intraclass variation of the query images. The experimental results on the AR database, Extended Yale B database, CMU-PIE database and LFW database show that the proposed method is robust to intra-class variations in FR with SSPP, and outperforms the state-of-art approaches.
基金supported by National Natural Science Foundation of China(60802069,61273270)the Fundamental Research Funds for the Central Universities of China+1 种基金Natural Science Foundation of Guangdong Province(2014A030313173)Science and Technology Program of Guangzhou(2014Y2-00165,2014J4100114,2014J4100095)
文摘Most face recognition techniques have been successful in dealing with high-resolution(HR) frontal face images. However, real-world face recognition systems are often confronted with the low-resolution(LR) face images with pose and illumination variations. This is a very challenging issue, especially under the constraint of using only a single gallery image per person.To address the problem, we propose a novel approach called coupled kernel-based enhanced discriminant analysis(CKEDA).CKEDA aims to simultaneously project the features from LR non-frontal probe images and HR frontal gallery ones into a common space where discrimination property is maximized.There are four advantages of the proposed approach: 1) by using the appropriate kernel function, the data becomes linearly separable, which is beneficial for recognition; 2) inspired by linear discriminant analysis(LDA), we integrate multiple discriminant factors into our objective function to enhance the discrimination property; 3) we use the gallery extended trick to improve the recognition performance for a single gallery image per person problem; 4) our approach can address the problem of matching LR non-frontal probe images with HR frontal gallery images,which is difficult for most existing face recognition techniques.Experimental evaluation on the multi-PIE dataset signifies highly competitive performance of our algorithm.
文摘It has been demonstrated that the linear discriminant analysis (LDA) is an effective approach in face recognition tasks. However, due to the high dimensionality of an image space, many LDA based approaches first use the principal component analysis (PCA) to project an image into a lower dimensional space, then perform the LDA transform to extract discriminant feature. But some useful discriminant information to the following LDA transform will be lost in the PCA step. To overcome these defects, a face recognition method based on the discrete cosine transform (DCT) and the LDA is proposed. First the DCT is used to achieve dimension reduction, then LDA transform is performed on the lower space to extract features. Two face databases are used to test our method and the correct recognition rates of 97.5% and 96.0% are obtained respectively. The performance of the proposed method is compared with that of the PCA+ LDA method and the results show that the method proposed outperforms the PCA+ LDA method.
基金This work was supported by Natural Science Foundation of Huazhong University of Science and Technology of PRC(No.2007Q006B).
文摘In order to distinguish faces of various angles during face recognition, an algorithm of the combination of approximate dynamic programming (ADP) called action dependent heuristic dynamic programming (ADHDP) and particle swarm optimization (PSO) is presented. ADP is used for dynamically changing the values of the PSO parameters. During the process of face recognition, the discrete cosine transformation (DCT) is first introduced to reduce negative effects. Then, Karhunen-Loeve (K-L) transformation can be used to compress images and decrease data dimensions. According to principal component analysis (PCA), the main parts of vectors are extracted for data representation. Finally, radial basis function (RBF) neural network is trained to recognize various faces. The training of RBF neural network is exploited by ADP-PSO. In terms of ORL Face Database, the experimental result gives a clear view of its accurate efficiency.
基金supported by the National Natural Science Foundation of China(6137901061772421)
文摘Face recognition based on few training samples is a challenging task. In daily applications, sufficient training samples may not be obtained and most of the gained training samples are in various illuminations and poses. Non-sufficient training samples could not effectively express various facial conditions, so the improvement of the face recognition rate under the non-sufficient training samples condition becomes a laborious mission. In our work, the facial pose pre-recognition(FPPR) model and the dualdictionary sparse representation classification(DD-SRC) are proposed for face recognition. The FPPR model is based on the facial geometric characteristic and machine learning, dividing a testing sample into full-face and profile. Different poses in a single dictionary are influenced by each other, which leads to a low face recognition rate. The DD-SRC contains two dictionaries, full-face dictionary and profile dictionary, and is able to reduce the interference. After FPPR, the sample is processed by the DD-SRC to find the most similar one in training samples. The experimental results show the performance of the proposed algorithm on olivetti research laboratory(ORL) and face recognition technology(FERET) databases, and also reflect comparisons with SRC, linear regression classification(LRC), and two-phase test sample sparse representation(TPTSSR).
文摘An improved face recognition method is proposed based on principal component analysis (PCA) compounded with genetic algorithm (GA), named as genetic based principal component analysis (GPCA). Initially the eigenspace is created with eigenvalues and eigenvectors. From this space, the eigenfaces are constructed, and the most relevant eigenfaees have been selected using GPCA. With these eigenfaees, the input images are classified based on Euclidian distance. The proposed method was tested on ORL (Olivetti Research Labs) face database. Experimental results on this database demonstrate that the effectiveness of the proposed method for face recognition has less misclassification in comparison with previous methods.
基金financial supports from the National Natural Science Foundation of China (No. 51134024)the National High Technology Research and Development Program of China (No. 2012AA062203)are gratefully acknowledged
文摘Since the coal mine in-pit personnel positioning system neither can effectively achieve the function to detect the uniqueness of in-pit coal-mine personnel nor can identify and eliminate violations in attendance management such as multiple cards for one person, and swiping one's cards by others in China at present. Therefore, the research introduces a uniqueness detection system and method for in-pit coal-mine personnel integrated into the in-pit coal mine personnel positioning system, establishing a system mode based on face recognition + recognition of personnel positioning card + release by automatic detection. Aiming at the facts that the in-pit personnel are wearing helmets and faces are prone to be stained during the face recognition, the study proposes the ideas that pre-process face images using the 2D-wavelet-transformation-based Mallat algorithm and extracts three face features: miner light, eyes and mouths, using the generalized symmetry transformation-based algorithm. This research carried out test with 40 clean face images with no helmets and 40 lightly-stained face images, and then compared with results with the one using the face feature extraction method based on grey-scale transformation and edge detection. The results show that the method described in the paper can detect accurately face features in the above-mentioned two cases, and the accuracy to detect face features is 97.5% in the case of wearing helmets and lightly-stained faces.
文摘Face recognition systems have enhanced human-computer interactions in the last ten years.However,the literature reveals that current techniques used for identifying or verifying faces are not immune to limitations.Principal Component Analysis-Support Vector Machine(PCA-SVM)and Principal Component Analysis-Artificial Neural Network(PCA-ANN)are among the relatively recent and powerful face analysis techniques.Compared to PCA-ANN,PCA-SVM has demonstrated generalization capabilities in many tasks,including the ability to recognize objects with small or large data samples.Apart from requiring a minimal number of parameters in face detection,PCA-SVM minimizes generalization errors and avoids overfitting problems better than PCA-ANN.PCA-SVM,however,is ineffective and inefficient in detecting human faces in cases in which there is poor lighting,long hair,or items covering the subject’s face.This study proposes a novel PCASVM-based model to overcome the recognition problem of PCA-ANN and enhance face detection.The experimental results indicate that the proposed model provides a better face recognition outcome than PCA-SVM.
基金supported by the National Natural Science Foundation of China under Grant No. 60973070
文摘An algorithm for face description and recognition based on multi-resolution with multi-scale local binary pattern (multi-LBP) features is proposed. The facial image pyramid is constructed and each facial image is divided into various regions from which partial and holistic local binary patter (LBP) histograms are extracted. All LBP features of each image are concatenated to a single LBP eigenvector with different resolutions. The dimensionaUty of LBP features is then reduced by a local margin alignment (LMA) algorithm based on manifold, which can preserve the between-class variance. Support vector machine (SVM) is applied to classify facial images. Extensive experiments on ORL and CMU face databases clearly show the superiority of the proposed scheme over some existed algorithms, especially on the robustness of the method against different facial expressions and postures of the subjects.
文摘Over the past few decades,face recognition has become the most effective biometric technique in recognizing people’s identity,as it is widely used in many areas of our daily lives.However,it is a challenging technique since facial images vary in rotations,expressions,and illuminations.To minimize the impact of these challenges,exploiting information from various feature extraction methods is recommended since one of the most critical tasks in face recognition system is the extraction of facial features.Therefore,this paper presents a new approach to face recognition based on the fusion of Gabor-based feature extraction,Fast Independent Component Analysis(FastICA),and Linear Discriminant Analysis(LDA).In the presented method,first,face images are transformed to grayscale and resized to have a uniform size.After that,facial features are extracted from the aligned face image using Gabor,FastICA,and LDA methods.Finally,the nearest distance classifier is utilized to recognize the identity of the individuals.Here,the performance of six distance classifiers,namely Euclidean,Cosine,Bray-Curtis,Mahalanobis,Correlation,and Manhattan,are investigated.Experimental results revealed that the presented method attains a higher rank-one recognition rate compared to the recent approaches in the literature on four benchmarked face datasets:ORL,GT,FEI,and Yale.Moreover,it showed that the proposed method not only helps in better extracting the features but also in improving the overall efficiency of the facial recognition system.
文摘In order to solve the problem that real-time face recognition is susceptible to illumination changes,this paper proposes a face recognition method that combines Local Binary Patterns(LBP)and Embedded Hidden Markov Model(EHMM).Face recognition method.The method firstly performs LBP preprocessing on the input face image,then extracts the feature vector,and finally sends the extracted feature observation vector to the EHMM for training or recognition.Experiments on multiple face databases show that the proposed algorithm is robust to illumination and improves recognition rate.
文摘This research presents an improved real-time face recognition system at a low<span><span><span style="font-family:" color:red;"=""> </span></span></span><span style="font-family:Verdana;"><span style="font-family:Verdana;"><span style="font-family:Verdana;">resolution of 15 pixels with pose and emotion and resolution variations. We have designed our datasets named LRD200 and LRD100, which have been used for training and classification. The face detection part uses the Viola-Jones algorithm, and the face recognition part receives the face image from the face detection part to process it using the Local Binary Pattern Histogram (LBPH) algorithm with preprocessing using contrast limited adaptive histogram equalization (CLAHE) and face alignment. The face database in this system can be updated via our custom-built standalone android app and automatic restarting of the training and recognition process with an updated database. Using our proposed algorithm, a real-time face recognition accuracy of 78.40% at 15</span></span></span><span><span><span style="font-family:;" "=""> </span></span></span><span style="font-family:Verdana;"><span style="font-family:Verdana;"><span style="font-family:Verdana;">px and 98.05% at 45</span></span></span><span><span><span style="font-family:;" "=""> </span></span></span><span style="font-family:Verdana;"><span style="font-family:Verdana;"><span style="font-family:Verdana;">px have been achieved using the LRD200 database containing 200 images per person. With 100 images per person in the database (LRD100) the achieved accuracies are 60.60% at 15</span></span></span><span><span><span style="font-family:;" "=""> </span></span></span><span style="font-family:Verdana;"><span style="font-family:Verdana;"><span style="font-family:Verdana;">px and 95% at 45</span></span></span><span><span><span style="font-family:;" "=""> </span></span></span><span style="font-family:Verdana;"><span style="font-family:Verdana;"><span style="font-family:Verdana;">px respectively. A facial deflection of about 30</span></span></span><span><span><span><span><span style="color:#4F4F4F;font-family:-apple-system, " font-size:16px;white-space:normal;background-color:#ffffff;"="">°</span></span><span> on either side from the front face showed an average face recognition precision of 72.25%-81.85%. This face recognition system can be employed for law enforcement purposes, where the surveillance camera captures a low-resolution image because of the distance of a person from the camera. It can also be used as a surveillance system in airports, bus stations, etc., to reduce the risk of possible criminal threats.</span></span></span></span>
文摘Time lapse, characteristic of aging, is a complex process that affects the reliability and security of biometric face recognition systems. This paper reports the novel use and effectiveness of deep learning, in general, and convolutional neural networks (CNN), in particular, for automatic rather than hand-crafted feature extraction for robust face recognition across time lapse. A CNN architecture using the VGG-Face deep (neural network) learning is found to produce highly discriminative and interoperable features that are robust to aging variations even across a mix of biometric datasets. The features extracted show high inter-class and low intra-class variability leading to low generalization errors on aging datasets using ensembles of subspace discriminant classifiers. The classification results for the all-encompassing authentication methods proposed on the challenging FG-NET and MORPH datasets are competitive with state-of-the-art methods including commercial face recognition engines and are richer in functionality and interoperability than existing methods as it handles mixed biometric datasets, e.g., FG-NET and MORPH.
文摘A face recognition scheme is proposed, wherein a face image is preprocessed by pixel averaging and energy normalizing to reduce data dimension and brightness variation effect, followed by the Fourier transform to estimate the spectrum of the preprocessed image. The principal component analysis is conducted on the spectra of a face image to obtain eigen features. Combining eigen features with a Parzen classifier, experiments are taken on the ORL face database.