Background Several face detection and recogni tion methods have been proposed in the past decades that have excellent performance.The conventional face recognition pipeline comprises the following:(1)face detection,(2...Background Several face detection and recogni tion methods have been proposed in the past decades that have excellent performance.The conventional face recognition pipeline comprises the following:(1)face detection,(2)face alignment,(3)feature extraction,and(4)similarity,which are independent of each other.The separate facial analysis stages lead to redundant model calculations,and are difficult for use in end-to-end training.Methods In this paper,we propose a novel end-to-end trainable convolutional network framework for face detection and recognition,in which a geometric transformation matrix is directly learned to align the faces rather than predicting the facial landmarks.In the training stage,our single CNN model is supervised only by face bounding boxes and personal identities,which are publicly available from WIDER FACE and CASIA-WebFace datasets.Our model is tested on Face Detection Dataset and Benchmark(FDDB)and Labeled Face in the Wild(LFW)datasets.Results The results show 89.24%recall for face detection tasks and 98.63%accura cy for face recognition tasks.展开更多
Since the coal mine in-pit personnel positioning system neither can effectively achieve the function to detect the uniqueness of in-pit coal-mine personnel nor can identify and eliminate violations in attendance manag...Since the coal mine in-pit personnel positioning system neither can effectively achieve the function to detect the uniqueness of in-pit coal-mine personnel nor can identify and eliminate violations in attendance management such as multiple cards for one person, and swiping one's cards by others in China at present. Therefore, the research introduces a uniqueness detection system and method for in-pit coal-mine personnel integrated into the in-pit coal mine personnel positioning system, establishing a system mode based on face recognition + recognition of personnel positioning card + release by automatic detection. Aiming at the facts that the in-pit personnel are wearing helmets and faces are prone to be stained during the face recognition, the study proposes the ideas that pre-process face images using the 2D-wavelet-transformation-based Mallat algorithm and extracts three face features: miner light, eyes and mouths, using the generalized symmetry transformation-based algorithm. This research carried out test with 40 clean face images with no helmets and 40 lightly-stained face images, and then compared with results with the one using the face feature extraction method based on grey-scale transformation and edge detection. The results show that the method described in the paper can detect accurately face features in the above-mentioned two cases, and the accuracy to detect face features is 97.5% in the case of wearing helmets and lightly-stained faces.展开更多
In this paper human face machine identification is experienced using optical correlation techniques in spatial frequency domain. This approach is tested on ORL dataset of faces which includes face images of 40 subject...In this paper human face machine identification is experienced using optical correlation techniques in spatial frequency domain. This approach is tested on ORL dataset of faces which includes face images of 40 subjects, each in 10 different positions. The examined optical setup relies on optical correlation based on developing optical Vanderlugt filters and its basics are described in this article. With the limitation of face database of 40 persons, the recognition is examined successfully with nearly 100% of accuracy in matching the input images with their respective Vanderlugt synthesized filters. Software simulation is implemented by using MATLAB for face identification.展开更多
Face recognition provides a natural visual interface for human computer interaction (HCI) applications. The process of face recognition, however, is inhibited by variations in the appearance of face images caused by...Face recognition provides a natural visual interface for human computer interaction (HCI) applications. The process of face recognition, however, is inhibited by variations in the appearance of face images caused by changes in lighting, expression, viewpoint, aging and introduction of occlusion. Although various algorithms have been presented for face recognition, face recognition is still a very challenging topic. A novel approach of real time face recognition for HCI is proposed in the paper. In view of the limits of the popular approaches to foreground segmentation, wavelet multi-scale transform based background subtraction is developed to extract foreground objects. The optimal selection of the threshold is automatically determined, which does not require any complex supervised training or manual experimental calibration. A robust real time face recognition algorithm is presented, which combines the projection matrixes without iteration and kernel Fisher discriminant analysis (KFDA) to overcome some difficulties existing in the real face recognition. Superior performance of the proposed algorithm is demonstrated by comparing with other algorithms through experiments. The proposed algorithm can also be applied to the video image sequences of natural HCI.展开更多
This research presents an improved real-time face recognition system at a low<span><span><span style="font-family:" color:red;"=""> </span></span></span><...This research presents an improved real-time face recognition system at a low<span><span><span style="font-family:" color:red;"=""> </span></span></span><span style="font-family:Verdana;"><span style="font-family:Verdana;"><span style="font-family:Verdana;">resolution of 15 pixels with pose and emotion and resolution variations. We have designed our datasets named LRD200 and LRD100, which have been used for training and classification. The face detection part uses the Viola-Jones algorithm, and the face recognition part receives the face image from the face detection part to process it using the Local Binary Pattern Histogram (LBPH) algorithm with preprocessing using contrast limited adaptive histogram equalization (CLAHE) and face alignment. The face database in this system can be updated via our custom-built standalone android app and automatic restarting of the training and recognition process with an updated database. Using our proposed algorithm, a real-time face recognition accuracy of 78.40% at 15</span></span></span><span><span><span style="font-family:;" "=""> </span></span></span><span style="font-family:Verdana;"><span style="font-family:Verdana;"><span style="font-family:Verdana;">px and 98.05% at 45</span></span></span><span><span><span style="font-family:;" "=""> </span></span></span><span style="font-family:Verdana;"><span style="font-family:Verdana;"><span style="font-family:Verdana;">px have been achieved using the LRD200 database containing 200 images per person. With 100 images per person in the database (LRD100) the achieved accuracies are 60.60% at 15</span></span></span><span><span><span style="font-family:;" "=""> </span></span></span><span style="font-family:Verdana;"><span style="font-family:Verdana;"><span style="font-family:Verdana;">px and 95% at 45</span></span></span><span><span><span style="font-family:;" "=""> </span></span></span><span style="font-family:Verdana;"><span style="font-family:Verdana;"><span style="font-family:Verdana;">px respectively. A facial deflection of about 30</span></span></span><span><span><span><span><span style="color:#4F4F4F;font-family:-apple-system, " font-size:16px;white-space:normal;background-color:#ffffff;"="">°</span></span><span> on either side from the front face showed an average face recognition precision of 72.25%-81.85%. This face recognition system can be employed for law enforcement purposes, where the surveillance camera captures a low-resolution image because of the distance of a person from the camera. It can also be used as a surveillance system in airports, bus stations, etc., to reduce the risk of possible criminal threats.</span></span></span></span>展开更多
This paper provides efficient and robust algorithms for real-time face detection and recognition in complex backgrounds. The algorithms are implemented using a series of signal processing methods including Ada Boost, ...This paper provides efficient and robust algorithms for real-time face detection and recognition in complex backgrounds. The algorithms are implemented using a series of signal processing methods including Ada Boost, cascade classifier, Local Binary Pattern (LBP), Haar-like feature, facial image pre-processing and Principal Component Analysis (PCA). The Ada Boost algorithm is implemented in a cascade classifier to train the face and eye detectors with robust detection accuracy. The LBP descriptor is utilized to extract facial features for fast face detection. The eye detection algorithm reduces the false face detection rate. The detected facial image is then processed to correct the orientation and increase the contrast, therefore, maintains high facial recognition accuracy. Finally, the PCA algorithm is used to recognize faces efficiently. Large databases with faces and non-faces images are used to train and validate face detection and facial recognition algorithms. The algorithms achieve an overall true-positive rate of 98.8% for face detection and 99.2% for correct facial recognition.展开更多
Fisherfaces algorithm is a popular method for face recognition.However,there exist some unstable com- ponents that degrade recognition performance.In this paper,we propose a method based on detecting reliable com- pon...Fisherfaces algorithm is a popular method for face recognition.However,there exist some unstable com- ponents that degrade recognition performance.In this paper,we propose a method based on detecting reliable com- ponents to overcome the problem and introduce it to 3D face recognition.The reliable components are detected within the binary feature vector,which is generated from the Fisherfaces feature vector based on statistical properties,and is used for 3D face recognition as the final feature vector.Experimental results show that the reliable components fea- ture vector is much more effective than the Fisherfaces feature vector for face recognition.展开更多
人脸识别技术广泛应用于考勤管理、移动支付等智慧建设中。伴随着常态化的口罩干扰,传统人脸识别算法已无法满足实际应用需求,为此,本文利用深度学习模型SSD以及FaceNet模型对人脸识别系统展开设计。首先,为消除现有数据集中亚洲人脸占...人脸识别技术广泛应用于考勤管理、移动支付等智慧建设中。伴随着常态化的口罩干扰,传统人脸识别算法已无法满足实际应用需求,为此,本文利用深度学习模型SSD以及FaceNet模型对人脸识别系统展开设计。首先,为消除现有数据集中亚洲人脸占比小造成的类内间距变化差距不明显的问题,在CAS-IA Web Face公开数据集的基础上对亚洲人脸数据进行扩充;其次,为解决不同口罩样式对特征提取的干扰,使用SSD人脸检测模型与DLIB人脸关键点检测模型提取人脸关键点,并利用人脸关键点与口罩的空间位置关系,额外随机生成不同的口罩人脸,组成混合数据集;最后,在混合数据集上进行模型训练并将训练好的模型移植到人脸识别系统中,进行检测速度与识别精度验证。实验结果表明,系统的实时识别速度达20 fps以上,人脸识别模型准确率在构建的混合数据集中达到97.1%,在随机抽取的部分LFW数据集验证的准确率达99.7%,故而该系统可满足实际应用需求,在一定程度上提高人脸识别的鲁棒性与准确性。展开更多
The automatic detection of faces is a very important problem. The effectiveness of biometric authentication based on face mainly depends on the method used to locate the face in the image. This paper presents a hybrid...The automatic detection of faces is a very important problem. The effectiveness of biometric authentication based on face mainly depends on the method used to locate the face in the image. This paper presents a hybrid system for faces detection in unconstrained cases in which the illumination, pose, occlusion, and size of the face are uncontrolled. To do this, the new method of detection proposed in this paper is based primarily on a technique of automatic learning by using the decision of three neural networks, a technique of energy compaction by using the discrete cosine transform, and a technique of segmentation by the color of human skin. A whole of pictures (faces and no faces) are transformed to vectors of data which will be used for learning the neural networks to separate between the two classes. Discrete cosine transform is used to reduce the dimension of the vectors, to eliminate the redundancies of information, and to store only the useful information in a minimum number of coefficients while the segmentation is used to reduce the space of research in the image. The experimental results have shown that this hybridization of methods will give a very significant improvement of the rate of the recognition, quality of detection, and the time of execution.展开更多
Optical flow estimation in human facial video,which provides 2D correspondences between adjacent frames,is a fundamental pre-processing step for many applications,like facial expression capture and recognition.However...Optical flow estimation in human facial video,which provides 2D correspondences between adjacent frames,is a fundamental pre-processing step for many applications,like facial expression capture and recognition.However,it is quite challenging as human facial images contain large areas of similar textures,rich expressions,and large rotations.These characteristics also result in the scarcity of large,annotated realworld datasets.We propose a robust and accurate method to learn facial optical flow in a self-supervised manner.Specifically,we utilize various shape priors,including face depth,landmarks,and parsing,to guide the self-supervised learning task via a differentiable nonrigid registration framework.Extensive experiments demonstrate that our method achieves remarkable improvements for facial optical flow estimation in the presence of significant expressions and large rotations.展开更多
This article proposes a feature extraction method for an integrated face tracking and facial expression recognition in real time video. The method proposed by Viola and Jones [1] is used to detect the face region in t...This article proposes a feature extraction method for an integrated face tracking and facial expression recognition in real time video. The method proposed by Viola and Jones [1] is used to detect the face region in the first frame of the video. A rectangular bounding box is fitted over for the face region and the detected face is tracked in the successive frames using the cascaded Support vector machine (SVM) and cascaded Radial basis function neural network (RBFNN). The haar-like features are extracted from the detected face region and they are used to create a cascaded SVM and RBFNN classifiers. Each stage of the SVM classifier and RBFNN classifier rejects the non-face regions and pass the face regions to the next stage in the cascade thereby efficiently tracking the face. The performance of tracking is evaluated using one hour video data. The performance of the cascaded SVM is compared with the cascaded RBFNN. The experiment results show that the proposed cascaded SVM classifier method gives better performance over the RBFNN and also the methods described in the literature using single SVM classifier [2]. While the face is being tracked, features are extracted from the mouth region for expression recognition. The features are modelled using a multi-class SVM. The SVM finds an optimal hyperplane to distinguish different facial expressions with an accuracy of 96.0%.展开更多
During Covid pandemic,many individuals are suffering from suicidal ideation in the world.Social distancing and quarantining,affects the patient emotionally.Affective computing is the study of recognizing human feeling...During Covid pandemic,many individuals are suffering from suicidal ideation in the world.Social distancing and quarantining,affects the patient emotionally.Affective computing is the study of recognizing human feelings and emotions.This technology can be used effectively during pandemic for facial expression recognition which automatically extracts the features from the human face.Monitoring system plays a very important role to detect the patient condition and to recognize the patterns of expression from the safest distance.In this paper,a new method is proposed for emotion recognition and suicide ideation detection in COVID patients.This helps to alert the nurse,when patient emotion is fear,cry or sad.The research presented in this paper has introduced Image Processing technology for emotional analysis of patients using Machine learning algorithm.The proposed Convolution Neural Networks(CNN)architecture with DnCNN preprocessing enhances the performance of recognition.The system can analyze the mood of patients either in real time or in the form of video files from CCTV cameras.The proposed method accuracy is more when compared to other methods.It detects the chances of suicide attempt based on stress level and emotional recognition.展开更多
Heart rate is an important vital characteristic which indicates physical and mental health status.Typically heart rate measurement instruments require direct contact with the skin which is time-consuming and costly.Th...Heart rate is an important vital characteristic which indicates physical and mental health status.Typically heart rate measurement instruments require direct contact with the skin which is time-consuming and costly.Therefore,the study of non-contact heart rate measurement methods is of great importance.Based on the principles of photoelectric volumetric tracing,we use a computer device and camera to capture facial images,accurately detect face regions,and to detect multiple facial images using a multi-target tracking algorithm.Then after the regional segmentation of the facial image,the signal acquisition of the region of interest is further resolved.Finally,frequency detection of the collected Photo-plethysmography(PPG)and Electrocardiography(ECG)signals is completed with peak detection,Fourier analysis,and a Waveletfilter.The experimental results show that the subject’s heart rate can be detected quickly and accurately even when monitoring multiple facial targets simultaneously.展开更多
Effective strategies to control COVID-19 pandemic need high attention to mitigate negatively impacted communal health and global economy,with the brim-full horizon yet to unfold.In the absence of effective antiviral a...Effective strategies to control COVID-19 pandemic need high attention to mitigate negatively impacted communal health and global economy,with the brim-full horizon yet to unfold.In the absence of effective antiviral and limited medical resources,many measures are recommended by WHO to control the infection rate and avoid exhausting the limited medical resources.Wearing mask is among the non-pharmaceutical intervention measures that can be used as barrier to primary route of SARS-CoV2 droplets expelled by presymptomatic or asymptomatic individuals.Regardless of discourse on medical resources and diversities in masks,all countries are mandating coverings over nose and mouth in public areas.Towards contribution of public health,the aim of the paper is to devise a real-time technique that can efficiently detect non mask faces in public and thus enforce to wear mask.The proposed technique is ensemble of one stage and two stage detectors to achieve low inference time and high accuracy.We took ResNet50 as a baseline model and applied the concept of transfer learning to fuse high level semantic information in multiple feature maps.In addition,we also propose a bounding box transformation to improve localization performance during mask detection.The experiments are conducted with three popular baseline models namely ResNet50,AlexNet and MobileNet.We explored the possibility of these models to plug-in with the proposed model,so that highly accurate results can be achieved in less inference time.It is observed that the proposed technique can achieve high accuracy(98.2%)when implemented with ResNet50.Besides,the proposed model can generate 11.07%and 6.44%higher precision and recall respectively in mask detection when compared to RetinaFaceMask detector.展开更多
文摘Background Several face detection and recogni tion methods have been proposed in the past decades that have excellent performance.The conventional face recognition pipeline comprises the following:(1)face detection,(2)face alignment,(3)feature extraction,and(4)similarity,which are independent of each other.The separate facial analysis stages lead to redundant model calculations,and are difficult for use in end-to-end training.Methods In this paper,we propose a novel end-to-end trainable convolutional network framework for face detection and recognition,in which a geometric transformation matrix is directly learned to align the faces rather than predicting the facial landmarks.In the training stage,our single CNN model is supervised only by face bounding boxes and personal identities,which are publicly available from WIDER FACE and CASIA-WebFace datasets.Our model is tested on Face Detection Dataset and Benchmark(FDDB)and Labeled Face in the Wild(LFW)datasets.Results The results show 89.24%recall for face detection tasks and 98.63%accura cy for face recognition tasks.
基金financial supports from the National Natural Science Foundation of China (No. 51134024)the National High Technology Research and Development Program of China (No. 2012AA062203)are gratefully acknowledged
文摘Since the coal mine in-pit personnel positioning system neither can effectively achieve the function to detect the uniqueness of in-pit coal-mine personnel nor can identify and eliminate violations in attendance management such as multiple cards for one person, and swiping one's cards by others in China at present. Therefore, the research introduces a uniqueness detection system and method for in-pit coal-mine personnel integrated into the in-pit coal mine personnel positioning system, establishing a system mode based on face recognition + recognition of personnel positioning card + release by automatic detection. Aiming at the facts that the in-pit personnel are wearing helmets and faces are prone to be stained during the face recognition, the study proposes the ideas that pre-process face images using the 2D-wavelet-transformation-based Mallat algorithm and extracts three face features: miner light, eyes and mouths, using the generalized symmetry transformation-based algorithm. This research carried out test with 40 clean face images with no helmets and 40 lightly-stained face images, and then compared with results with the one using the face feature extraction method based on grey-scale transformation and edge detection. The results show that the method described in the paper can detect accurately face features in the above-mentioned two cases, and the accuracy to detect face features is 97.5% in the case of wearing helmets and lightly-stained faces.
文摘In this paper human face machine identification is experienced using optical correlation techniques in spatial frequency domain. This approach is tested on ORL dataset of faces which includes face images of 40 subjects, each in 10 different positions. The examined optical setup relies on optical correlation based on developing optical Vanderlugt filters and its basics are described in this article. With the limitation of face database of 40 persons, the recognition is examined successfully with nearly 100% of accuracy in matching the input images with their respective Vanderlugt synthesized filters. Software simulation is implemented by using MATLAB for face identification.
基金supported by the National Natural Science Foundation of China (Grant No.60872117)the Leading Academic Discipline Project of Shanghai Municipal Education Commission (Grant No.J50104)
文摘Face recognition provides a natural visual interface for human computer interaction (HCI) applications. The process of face recognition, however, is inhibited by variations in the appearance of face images caused by changes in lighting, expression, viewpoint, aging and introduction of occlusion. Although various algorithms have been presented for face recognition, face recognition is still a very challenging topic. A novel approach of real time face recognition for HCI is proposed in the paper. In view of the limits of the popular approaches to foreground segmentation, wavelet multi-scale transform based background subtraction is developed to extract foreground objects. The optimal selection of the threshold is automatically determined, which does not require any complex supervised training or manual experimental calibration. A robust real time face recognition algorithm is presented, which combines the projection matrixes without iteration and kernel Fisher discriminant analysis (KFDA) to overcome some difficulties existing in the real face recognition. Superior performance of the proposed algorithm is demonstrated by comparing with other algorithms through experiments. The proposed algorithm can also be applied to the video image sequences of natural HCI.
文摘This research presents an improved real-time face recognition system at a low<span><span><span style="font-family:" color:red;"=""> </span></span></span><span style="font-family:Verdana;"><span style="font-family:Verdana;"><span style="font-family:Verdana;">resolution of 15 pixels with pose and emotion and resolution variations. We have designed our datasets named LRD200 and LRD100, which have been used for training and classification. The face detection part uses the Viola-Jones algorithm, and the face recognition part receives the face image from the face detection part to process it using the Local Binary Pattern Histogram (LBPH) algorithm with preprocessing using contrast limited adaptive histogram equalization (CLAHE) and face alignment. The face database in this system can be updated via our custom-built standalone android app and automatic restarting of the training and recognition process with an updated database. Using our proposed algorithm, a real-time face recognition accuracy of 78.40% at 15</span></span></span><span><span><span style="font-family:;" "=""> </span></span></span><span style="font-family:Verdana;"><span style="font-family:Verdana;"><span style="font-family:Verdana;">px and 98.05% at 45</span></span></span><span><span><span style="font-family:;" "=""> </span></span></span><span style="font-family:Verdana;"><span style="font-family:Verdana;"><span style="font-family:Verdana;">px have been achieved using the LRD200 database containing 200 images per person. With 100 images per person in the database (LRD100) the achieved accuracies are 60.60% at 15</span></span></span><span><span><span style="font-family:;" "=""> </span></span></span><span style="font-family:Verdana;"><span style="font-family:Verdana;"><span style="font-family:Verdana;">px and 95% at 45</span></span></span><span><span><span style="font-family:;" "=""> </span></span></span><span style="font-family:Verdana;"><span style="font-family:Verdana;"><span style="font-family:Verdana;">px respectively. A facial deflection of about 30</span></span></span><span><span><span><span><span style="color:#4F4F4F;font-family:-apple-system, " font-size:16px;white-space:normal;background-color:#ffffff;"="">°</span></span><span> on either side from the front face showed an average face recognition precision of 72.25%-81.85%. This face recognition system can be employed for law enforcement purposes, where the surveillance camera captures a low-resolution image because of the distance of a person from the camera. It can also be used as a surveillance system in airports, bus stations, etc., to reduce the risk of possible criminal threats.</span></span></span></span>
文摘This paper provides efficient and robust algorithms for real-time face detection and recognition in complex backgrounds. The algorithms are implemented using a series of signal processing methods including Ada Boost, cascade classifier, Local Binary Pattern (LBP), Haar-like feature, facial image pre-processing and Principal Component Analysis (PCA). The Ada Boost algorithm is implemented in a cascade classifier to train the face and eye detectors with robust detection accuracy. The LBP descriptor is utilized to extract facial features for fast face detection. The eye detection algorithm reduces the false face detection rate. The detected facial image is then processed to correct the orientation and increase the contrast, therefore, maintains high facial recognition accuracy. Finally, the PCA algorithm is used to recognize faces efficiently. Large databases with faces and non-faces images are used to train and validate face detection and facial recognition algorithms. The algorithms achieve an overall true-positive rate of 98.8% for face detection and 99.2% for correct facial recognition.
基金Supported by the National Natural Science Foundation of China(60671064)the Foundation for the Author of National Excellent Doctoral Dissertation of China(FANEDD-200238)+1 种基金the Foundation for the Excellent Youth of Heilongjiang Provincethe Program for New Century Excellent Talents in University(NCET-04-0330)
文摘Fisherfaces algorithm is a popular method for face recognition.However,there exist some unstable com- ponents that degrade recognition performance.In this paper,we propose a method based on detecting reliable com- ponents to overcome the problem and introduce it to 3D face recognition.The reliable components are detected within the binary feature vector,which is generated from the Fisherfaces feature vector based on statistical properties,and is used for 3D face recognition as the final feature vector.Experimental results show that the reliable components fea- ture vector is much more effective than the Fisherfaces feature vector for face recognition.
文摘人脸识别技术广泛应用于考勤管理、移动支付等智慧建设中。伴随着常态化的口罩干扰,传统人脸识别算法已无法满足实际应用需求,为此,本文利用深度学习模型SSD以及FaceNet模型对人脸识别系统展开设计。首先,为消除现有数据集中亚洲人脸占比小造成的类内间距变化差距不明显的问题,在CAS-IA Web Face公开数据集的基础上对亚洲人脸数据进行扩充;其次,为解决不同口罩样式对特征提取的干扰,使用SSD人脸检测模型与DLIB人脸关键点检测模型提取人脸关键点,并利用人脸关键点与口罩的空间位置关系,额外随机生成不同的口罩人脸,组成混合数据集;最后,在混合数据集上进行模型训练并将训练好的模型移植到人脸识别系统中,进行检测速度与识别精度验证。实验结果表明,系统的实时识别速度达20 fps以上,人脸识别模型准确率在构建的混合数据集中达到97.1%,在随机抽取的部分LFW数据集验证的准确率达99.7%,故而该系统可满足实际应用需求,在一定程度上提高人脸识别的鲁棒性与准确性。
基金supported by the Laboratory of Inverses Problems, Modeling, Information and Systems (PI:MIS), Department of Electronic and Telecommunication, University of 08 Mai 1945, Guelma, Algériathe Laboratory of Computer Research (LRI), Department of Computer Sciences, University of Badji Mokhtar, Annaba, Algéria
文摘The automatic detection of faces is a very important problem. The effectiveness of biometric authentication based on face mainly depends on the method used to locate the face in the image. This paper presents a hybrid system for faces detection in unconstrained cases in which the illumination, pose, occlusion, and size of the face are uncontrolled. To do this, the new method of detection proposed in this paper is based primarily on a technique of automatic learning by using the decision of three neural networks, a technique of energy compaction by using the discrete cosine transform, and a technique of segmentation by the color of human skin. A whole of pictures (faces and no faces) are transformed to vectors of data which will be used for learning the neural networks to separate between the two classes. Discrete cosine transform is used to reduce the dimension of the vectors, to eliminate the redundancies of information, and to store only the useful information in a minimum number of coefficients while the segmentation is used to reduce the space of research in the image. The experimental results have shown that this hybridization of methods will give a very significant improvement of the rate of the recognition, quality of detection, and the time of execution.
基金This work was supported by National Natural Science Foundation of China(No.62122071)the Youth Innovation Promotion Association CAS(No.2018495)+1 种基金the Fundamental Research Funds for the Central Universities(No.WK3470000021)through the Alibaba Innovation Research Program(AIR).
文摘Optical flow estimation in human facial video,which provides 2D correspondences between adjacent frames,is a fundamental pre-processing step for many applications,like facial expression capture and recognition.However,it is quite challenging as human facial images contain large areas of similar textures,rich expressions,and large rotations.These characteristics also result in the scarcity of large,annotated realworld datasets.We propose a robust and accurate method to learn facial optical flow in a self-supervised manner.Specifically,we utilize various shape priors,including face depth,landmarks,and parsing,to guide the self-supervised learning task via a differentiable nonrigid registration framework.Extensive experiments demonstrate that our method achieves remarkable improvements for facial optical flow estimation in the presence of significant expressions and large rotations.
文摘This article proposes a feature extraction method for an integrated face tracking and facial expression recognition in real time video. The method proposed by Viola and Jones [1] is used to detect the face region in the first frame of the video. A rectangular bounding box is fitted over for the face region and the detected face is tracked in the successive frames using the cascaded Support vector machine (SVM) and cascaded Radial basis function neural network (RBFNN). The haar-like features are extracted from the detected face region and they are used to create a cascaded SVM and RBFNN classifiers. Each stage of the SVM classifier and RBFNN classifier rejects the non-face regions and pass the face regions to the next stage in the cascade thereby efficiently tracking the face. The performance of tracking is evaluated using one hour video data. The performance of the cascaded SVM is compared with the cascaded RBFNN. The experiment results show that the proposed cascaded SVM classifier method gives better performance over the RBFNN and also the methods described in the literature using single SVM classifier [2]. While the face is being tracked, features are extracted from the mouth region for expression recognition. The features are modelled using a multi-class SVM. The SVM finds an optimal hyperplane to distinguish different facial expressions with an accuracy of 96.0%.
文摘During Covid pandemic,many individuals are suffering from suicidal ideation in the world.Social distancing and quarantining,affects the patient emotionally.Affective computing is the study of recognizing human feelings and emotions.This technology can be used effectively during pandemic for facial expression recognition which automatically extracts the features from the human face.Monitoring system plays a very important role to detect the patient condition and to recognize the patterns of expression from the safest distance.In this paper,a new method is proposed for emotion recognition and suicide ideation detection in COVID patients.This helps to alert the nurse,when patient emotion is fear,cry or sad.The research presented in this paper has introduced Image Processing technology for emotional analysis of patients using Machine learning algorithm.The proposed Convolution Neural Networks(CNN)architecture with DnCNN preprocessing enhances the performance of recognition.The system can analyze the mood of patients either in real time or in the form of video files from CCTV cameras.The proposed method accuracy is more when compared to other methods.It detects the chances of suicide attempt based on stress level and emotional recognition.
基金supported by the National Nature Science Foundation of China(Grant Number:61962010).
文摘Heart rate is an important vital characteristic which indicates physical and mental health status.Typically heart rate measurement instruments require direct contact with the skin which is time-consuming and costly.Therefore,the study of non-contact heart rate measurement methods is of great importance.Based on the principles of photoelectric volumetric tracing,we use a computer device and camera to capture facial images,accurately detect face regions,and to detect multiple facial images using a multi-target tracking algorithm.Then after the regional segmentation of the facial image,the signal acquisition of the region of interest is further resolved.Finally,frequency detection of the collected Photo-plethysmography(PPG)and Electrocardiography(ECG)signals is completed with peak detection,Fourier analysis,and a Waveletfilter.The experimental results show that the subject’s heart rate can be detected quickly and accurately even when monitoring multiple facial targets simultaneously.
文摘Effective strategies to control COVID-19 pandemic need high attention to mitigate negatively impacted communal health and global economy,with the brim-full horizon yet to unfold.In the absence of effective antiviral and limited medical resources,many measures are recommended by WHO to control the infection rate and avoid exhausting the limited medical resources.Wearing mask is among the non-pharmaceutical intervention measures that can be used as barrier to primary route of SARS-CoV2 droplets expelled by presymptomatic or asymptomatic individuals.Regardless of discourse on medical resources and diversities in masks,all countries are mandating coverings over nose and mouth in public areas.Towards contribution of public health,the aim of the paper is to devise a real-time technique that can efficiently detect non mask faces in public and thus enforce to wear mask.The proposed technique is ensemble of one stage and two stage detectors to achieve low inference time and high accuracy.We took ResNet50 as a baseline model and applied the concept of transfer learning to fuse high level semantic information in multiple feature maps.In addition,we also propose a bounding box transformation to improve localization performance during mask detection.The experiments are conducted with three popular baseline models namely ResNet50,AlexNet and MobileNet.We explored the possibility of these models to plug-in with the proposed model,so that highly accurate results can be achieved in less inference time.It is observed that the proposed technique can achieve high accuracy(98.2%)when implemented with ResNet50.Besides,the proposed model can generate 11.07%and 6.44%higher precision and recall respectively in mask detection when compared to RetinaFaceMask detector.