In this paper, we proposed a new approach for face recognition with robust to illumination variation. The improved performance to various lights in recognition is obtained by a novel combination of multicondition reli...In this paper, we proposed a new approach for face recognition with robust to illumination variation. The improved performance to various lights in recognition is obtained by a novel combination of multicondition relighting and optimal feature selection. Multi-condition relighting provides a "coarse" compensation for the variable illumination, and then the optimal feature selection further refines the compensation, and additionally offers the robustness to shadow and highlight, by deemphasizing the local mismatches caused by imprecise lighting compensation, shadow or highlight on recognition. For evaluation, two databases with various illumination mismatches have been used. The results have demonstrated the improved robustness of the new methods.展开更多
This study aimed to address the challenge of accurately and reliably detecting tomatoes in dense planting environments,a critical prerequisite for the automation implementation of robotic harvesting.However,the heavy ...This study aimed to address the challenge of accurately and reliably detecting tomatoes in dense planting environments,a critical prerequisite for the automation implementation of robotic harvesting.However,the heavy reliance on extensive manually annotated datasets for training deep learning models still poses significant limitations to their application in real-world agricultural production environments.To overcome these limitations,we employed domain adaptive learning approach combined with the YOLOv5 model to develop a novel tomato detection model called as TDA-YOLO(tomato detection domain adaptation).We designated the normal illumination scenes in dense planting environments as the source domain and utilized various other illumination scenes as the target domain.To construct bridge mechanism between source and target domains,neural preset for color style transfer is introduced to generate a pseudo-dataset,which served to deal with domain discrepancy.Furthermore,this study combines the semi-supervised learning method to enable the model to extract domain-invariant features more fully,and uses knowledge distillation to improve the model's ability to adapt to the target domain.Additionally,for purpose of promoting inference speed and low computational demand,the lightweight FasterNet network was integrated into the YOLOv5's C3 module,creating a modified C3_Faster module.The experimental results demonstrated that the proposed TDA-YOLO model significantly outperformed original YOLOv5s model,achieving a mAP(mean average precision)of 96.80%for tomato detection across diverse scenarios in dense planting environments,increasing by 7.19 percentage points;Compared with the latest YOLOv8 and YOLOv9,it is also 2.17 and 1.19 percentage points higher,respectively.The model's average detection time per image was an impressive 15 milliseconds,with a FLOPs(floating point operations per second)count of 13.8 G.After acceleration processing,the detection accuracy of the TDA-YOLO model on the Jetson Xavier NX development board is 90.95%,the mAP value is 91.35%,and the detection time of each image is 21 ms,which can still meet the requirements of real-time detection of tomatoes in dense planting environment.The experimental results show that the proposed TDA-YOLO model can accurately and quickly detect tomatoes in dense planting environment,and at the same time avoid the use of a large number of annotated data,which provides technical support for the development of automatic harvesting systems for tomatoes and other fruits.展开更多
A centre symmetric quadruple pattern-based illumination invariant measure(CSQPIM)is proposed to tackle severe illumination variation face recognition.First,the subtraction of the pixel pairs of the centre symmetric qu...A centre symmetric quadruple pattern-based illumination invariant measure(CSQPIM)is proposed to tackle severe illumination variation face recognition.First,the subtraction of the pixel pairs of the centre symmetric quadruple pattern(CSQP)is defined as the CSQPIM unit in the logarithm face local region,which may be positive or negative.The CSQPIM model is obtained by combining the positive and negative CSQPIM units.Then,the CSQPIM model can be used to generate several CSQPIM images by controlling the proportions of positive and negative CSQPIM units.The single CSQPIM image with the saturation function can be used to develop the CSQPIM-face.Multi CSQPIM images employ the extended sparse representation classification(ESRC)as the classifier,which can create the CSQPIM image-based classification(CSQPIMC).Furthermore,the CSQPIM model is integrated with the pre-trained deep learning(PDL)model to construct the CSQPIM-PDL model.Finally,the experimental results on the Extended Yale B,CMU PIE and Driver face databases indicate that the proposed methods are efficient for tackling severe illumination variations.展开更多
Local Binary Patterns (LBPs) have been highly used in texture classification <span style="font-family:Verdana;">for their robustness, their ease of implementation an</span><span style="fo...Local Binary Patterns (LBPs) have been highly used in texture classification <span style="font-family:Verdana;">for their robustness, their ease of implementation an</span><span style="font-family:Verdana;">d their low computational</span><span style="font-family:;" "=""> </span><span style="font-family:;" "=""><span style="font-family:Verdana;">cost. Initially designed to deal with gray level images, several methods based on them in the literature have been proposed for images having more than one spectral band. To achieve it, whether assumption using color information or combining spectral band two by two was done. Those methods use micro </span><span style="font-family:Verdana;">structures as texture features. In this paper, our goal was to design texture features which are relevant to color and multicomponent texture analysi</span><span style="font-family:Verdana;">s withou</span><span style="font-family:Verdana;">t any assumption.</span></span><span style="font-family:;" "=""> </span><span style="font-family:;" "=""><span style="font-family:Verdana;">Based on methods designed for gray scale images, we find the combination of micro and macro structures efficient for multispectral texture analysis. The experimentations were carried out on color images from Outex databases and multicomponent images from red blood cells captured using a multispectral microscope equipped with 13 LEDs ranging </span><span style="font-family:Verdana;">from 375 nm to 940 nm. In all achieved experimentations, our propos</span><span style="font-family:Verdana;">al presents the best classification scores compared to common multicomponent LBP methods.</span></span><span style="font-family:;" "=""> </span><span style="font-family:Verdana;">99.81%, 100.00%,</span><span style="font-family:;" "=""> </span><span style="font-family:Verdana;">99.07% and 97.67% are</span><span style="font-family:;" "=""> </span><span style="font-family:Verdana;">maximum scores obtained with our strategy respectively applied to images subject to rotation, blur, illumination variation and the multicomponent ones.</span>展开更多
A novel adaptive illumination normalization approach is proposed to eliminate the effects caused by illumination variations for face recognition. The proposed method divides an image into blocks and performs discrete ...A novel adaptive illumination normalization approach is proposed to eliminate the effects caused by illumination variations for face recognition. The proposed method divides an image into blocks and performs discrete cosine transform(DCT) in blocks independently in the logarithm domain. For each block-DCT coefficient except the direct current(DC) component, we take the illumination as main signal and take the reflectance as "noise". A data-driven and adaptive soft-thresholding denoising technique is employed in each block-DCT coefficient except the DC component. Illumination is estimated by applying the inverse DCT in the block-DCT coefficients, and the indirectly obtained reflectance can be used in further recognition task. Experimental results show that the proposed approach outperforms other existing methods. Moreover, the proposed method does not need any prior information, and none of the parameters can be determined by experience.展开更多
Classification of different weather conditions provides a first step support for outdoor scene modeling,which is a core component in many different applications of outdoor video analysis and computer vision.Features d...Classification of different weather conditions provides a first step support for outdoor scene modeling,which is a core component in many different applications of outdoor video analysis and computer vision.Features derived from intrinsic properties of the visual effects of different weather conditions contribute to successful classification.In this paper,features representing both the autocorrelation of pixel-wise intensities over time and the max directional length of rain streaks or snowflakes are proposed.Based on the autocorrelation of each pixel’s intensities over time,two temporal features are used for coarse classification of weather conditions according to their visual effects.On the other hand,features are extracted for fine classification of video clips with rain and snow.The classification results on 249 video clips associated with different weather conditions indicate the effectiveness of the extracted features,by using C-SVM as the classifier.展开更多
文摘In this paper, we proposed a new approach for face recognition with robust to illumination variation. The improved performance to various lights in recognition is obtained by a novel combination of multicondition relighting and optimal feature selection. Multi-condition relighting provides a "coarse" compensation for the variable illumination, and then the optimal feature selection further refines the compensation, and additionally offers the robustness to shadow and highlight, by deemphasizing the local mismatches caused by imprecise lighting compensation, shadow or highlight on recognition. For evaluation, two databases with various illumination mismatches have been used. The results have demonstrated the improved robustness of the new methods.
基金The National Natural Science Foundation of China (32371993)The Natural Science Research Key Project of Anhui Provincial University(2022AH040125&2023AH040135)The Key Research and Development Plan of Anhui Province (202204c06020022&2023n06020057)。
文摘This study aimed to address the challenge of accurately and reliably detecting tomatoes in dense planting environments,a critical prerequisite for the automation implementation of robotic harvesting.However,the heavy reliance on extensive manually annotated datasets for training deep learning models still poses significant limitations to their application in real-world agricultural production environments.To overcome these limitations,we employed domain adaptive learning approach combined with the YOLOv5 model to develop a novel tomato detection model called as TDA-YOLO(tomato detection domain adaptation).We designated the normal illumination scenes in dense planting environments as the source domain and utilized various other illumination scenes as the target domain.To construct bridge mechanism between source and target domains,neural preset for color style transfer is introduced to generate a pseudo-dataset,which served to deal with domain discrepancy.Furthermore,this study combines the semi-supervised learning method to enable the model to extract domain-invariant features more fully,and uses knowledge distillation to improve the model's ability to adapt to the target domain.Additionally,for purpose of promoting inference speed and low computational demand,the lightweight FasterNet network was integrated into the YOLOv5's C3 module,creating a modified C3_Faster module.The experimental results demonstrated that the proposed TDA-YOLO model significantly outperformed original YOLOv5s model,achieving a mAP(mean average precision)of 96.80%for tomato detection across diverse scenarios in dense planting environments,increasing by 7.19 percentage points;Compared with the latest YOLOv8 and YOLOv9,it is also 2.17 and 1.19 percentage points higher,respectively.The model's average detection time per image was an impressive 15 milliseconds,with a FLOPs(floating point operations per second)count of 13.8 G.After acceleration processing,the detection accuracy of the TDA-YOLO model on the Jetson Xavier NX development board is 90.95%,the mAP value is 91.35%,and the detection time of each image is 21 ms,which can still meet the requirements of real-time detection of tomatoes in dense planting environment.The experimental results show that the proposed TDA-YOLO model can accurately and quickly detect tomatoes in dense planting environment,and at the same time avoid the use of a large number of annotated data,which provides technical support for the development of automatic harvesting systems for tomatoes and other fruits.
基金The National Natural Science Foundation of China(No.61802203)the Natural Science Foundation of Jiangsu Province(No.BK20180761)+1 种基金China Postdoctoral Science Foundation(No.2019M651653)Postdoctoral Research Funding Program of Jiangsu Province(No.2019K124).
文摘A centre symmetric quadruple pattern-based illumination invariant measure(CSQPIM)is proposed to tackle severe illumination variation face recognition.First,the subtraction of the pixel pairs of the centre symmetric quadruple pattern(CSQP)is defined as the CSQPIM unit in the logarithm face local region,which may be positive or negative.The CSQPIM model is obtained by combining the positive and negative CSQPIM units.Then,the CSQPIM model can be used to generate several CSQPIM images by controlling the proportions of positive and negative CSQPIM units.The single CSQPIM image with the saturation function can be used to develop the CSQPIM-face.Multi CSQPIM images employ the extended sparse representation classification(ESRC)as the classifier,which can create the CSQPIM image-based classification(CSQPIMC).Furthermore,the CSQPIM model is integrated with the pre-trained deep learning(PDL)model to construct the CSQPIM-PDL model.Finally,the experimental results on the Extended Yale B,CMU PIE and Driver face databases indicate that the proposed methods are efficient for tackling severe illumination variations.
文摘Local Binary Patterns (LBPs) have been highly used in texture classification <span style="font-family:Verdana;">for their robustness, their ease of implementation an</span><span style="font-family:Verdana;">d their low computational</span><span style="font-family:;" "=""> </span><span style="font-family:;" "=""><span style="font-family:Verdana;">cost. Initially designed to deal with gray level images, several methods based on them in the literature have been proposed for images having more than one spectral band. To achieve it, whether assumption using color information or combining spectral band two by two was done. Those methods use micro </span><span style="font-family:Verdana;">structures as texture features. In this paper, our goal was to design texture features which are relevant to color and multicomponent texture analysi</span><span style="font-family:Verdana;">s withou</span><span style="font-family:Verdana;">t any assumption.</span></span><span style="font-family:;" "=""> </span><span style="font-family:;" "=""><span style="font-family:Verdana;">Based on methods designed for gray scale images, we find the combination of micro and macro structures efficient for multispectral texture analysis. The experimentations were carried out on color images from Outex databases and multicomponent images from red blood cells captured using a multispectral microscope equipped with 13 LEDs ranging </span><span style="font-family:Verdana;">from 375 nm to 940 nm. In all achieved experimentations, our propos</span><span style="font-family:Verdana;">al presents the best classification scores compared to common multicomponent LBP methods.</span></span><span style="font-family:;" "=""> </span><span style="font-family:Verdana;">99.81%, 100.00%,</span><span style="font-family:;" "=""> </span><span style="font-family:Verdana;">99.07% and 97.67% are</span><span style="font-family:;" "=""> </span><span style="font-family:Verdana;">maximum scores obtained with our strategy respectively applied to images subject to rotation, blur, illumination variation and the multicomponent ones.</span>
基金the Natural Science Foundation of Jiangsu Province(No.BK20150784)the Fund of Jiangsu Key Laboratory of Image and Video Understanding for Social Safety(Nanjing University of Science and Technology)(No.30920140122007)
文摘A novel adaptive illumination normalization approach is proposed to eliminate the effects caused by illumination variations for face recognition. The proposed method divides an image into blocks and performs discrete cosine transform(DCT) in blocks independently in the logarithm domain. For each block-DCT coefficient except the direct current(DC) component, we take the illumination as main signal and take the reflectance as "noise". A data-driven and adaptive soft-thresholding denoising technique is employed in each block-DCT coefficient except the DC component. Illumination is estimated by applying the inverse DCT in the block-DCT coefficients, and the indirectly obtained reflectance can be used in further recognition task. Experimental results show that the proposed approach outperforms other existing methods. Moreover, the proposed method does not need any prior information, and none of the parameters can be determined by experience.
基金This work was supported by the National Natural Science Foundation of China(Grant No.60702032)the Natural Scientific Research Innovation Foundation in Harbin Institute of Technology(No.HIT.NSRIF.2008.63)China Academy of Space Technology Innovation Foundation(No.CAST200814).
文摘Classification of different weather conditions provides a first step support for outdoor scene modeling,which is a core component in many different applications of outdoor video analysis and computer vision.Features derived from intrinsic properties of the visual effects of different weather conditions contribute to successful classification.In this paper,features representing both the autocorrelation of pixel-wise intensities over time and the max directional length of rain streaks or snowflakes are proposed.Based on the autocorrelation of each pixel’s intensities over time,two temporal features are used for coarse classification of weather conditions according to their visual effects.On the other hand,features are extracted for fine classification of video clips with rain and snow.The classification results on 249 video clips associated with different weather conditions indicate the effectiveness of the extracted features,by using C-SVM as the classifier.