Image-based face pose estimation tries to estimate the facial direction with 2D images.It provides important information for many face recognition applications.However,it is a difficult task due to complex conditions ...Image-based face pose estimation tries to estimate the facial direction with 2D images.It provides important information for many face recognition applications.However,it is a difficult task due to complex conditions and appearances.Deep learning method used in this field has the disadvantage of ignoring the natural structures of human faces.To solve this problem,a framework is proposed in this paper to estimate face poses with regression,which is based on deep learning and multi-modal feature loss(M2FL).Different from current loss functions using only a single type of features,the descriptive power was improved by combining multiple image features.To achieve it,hypergraph-based manifold regularization was applied.In this way,the loss of face pose estimation was reduced.Experimental results on commonly-used benchmark datasets demonstrate the performance of M2FL.展开更多
Face image analysis is one among several important cues in computer vision.Over the last five decades,methods for face analysis have received immense attention due to large scale applications in various face analysis ...Face image analysis is one among several important cues in computer vision.Over the last five decades,methods for face analysis have received immense attention due to large scale applications in various face analysis tasks.Face parsing strongly benefits various human face image analysis tasks inducing face pose estimation.In this paper we propose a 3D head pose estimation framework developed through a prior end to end deep face parsing model.We have developed an end to end face parts segmentation framework through deep convolutional neural networks(DCNNs).For training a deep face parts parsing model,we label face images for seven different classes,including eyes,brows,nose,hair,mouth,skin,and back.We extract features from gray scale images by using DCNNs.We train a classifier using the extracted features.We use the probabilistic classification method to produce gray scale images in the form of probability maps for each dense semantic class.We use a next stage of DCNNs and extract features from grayscale images created as probability maps during the segmentation phase.We assess the performance of our newly proposed model on four standard head pose datasets,including Pointing’04,Annotated Facial Landmarks in the Wild(AFLW),Boston University(BU),and ICT-3DHP,obtaining superior results as compared to previous results.展开更多
基金the National Natural Science Foundation of China(61871464 and 61836002)the Fujian Provincial Natural Science Foundation of China(2018J01573)+1 种基金the Foundation of Fujian Educational Committee(JAT160357)Distinguished Young Scientific Research Talents Plan in Universities of Fujian Province and the Program for New Century Excellent Talents in University of Fujian Province.
文摘Image-based face pose estimation tries to estimate the facial direction with 2D images.It provides important information for many face recognition applications.However,it is a difficult task due to complex conditions and appearances.Deep learning method used in this field has the disadvantage of ignoring the natural structures of human faces.To solve this problem,a framework is proposed in this paper to estimate face poses with regression,which is based on deep learning and multi-modal feature loss(M2FL).Different from current loss functions using only a single type of features,the descriptive power was improved by combining multiple image features.To achieve it,hypergraph-based manifold regularization was applied.In this way,the loss of face pose estimation was reduced.Experimental results on commonly-used benchmark datasets demonstrate the performance of M2FL.
基金Institute of Information&communications Technology Planning&Evaluation(IITP)grant funded by the Korea government(MSIT)(2020-0-01592)Basic Science Research Program through the National Research Foundation of Korea(NRF)funded by the Ministry of Education under Grant(2019R1F1A1058548)and Grant(2020R1G1A1013221).
文摘Face image analysis is one among several important cues in computer vision.Over the last five decades,methods for face analysis have received immense attention due to large scale applications in various face analysis tasks.Face parsing strongly benefits various human face image analysis tasks inducing face pose estimation.In this paper we propose a 3D head pose estimation framework developed through a prior end to end deep face parsing model.We have developed an end to end face parts segmentation framework through deep convolutional neural networks(DCNNs).For training a deep face parts parsing model,we label face images for seven different classes,including eyes,brows,nose,hair,mouth,skin,and back.We extract features from gray scale images by using DCNNs.We train a classifier using the extracted features.We use the probabilistic classification method to produce gray scale images in the form of probability maps for each dense semantic class.We use a next stage of DCNNs and extract features from grayscale images created as probability maps during the segmentation phase.We assess the performance of our newly proposed model on four standard head pose datasets,including Pointing’04,Annotated Facial Landmarks in the Wild(AFLW),Boston University(BU),and ICT-3DHP,obtaining superior results as compared to previous results.