This paper presents two systems for recognizing static signs (digits) from American Sign Language (ASL). These systems avoid the use color marks, or gloves, using instead, low-pass and high-pass filters in space and f...This paper presents two systems for recognizing static signs (digits) from American Sign Language (ASL). These systems avoid the use color marks, or gloves, using instead, low-pass and high-pass filters in space and frequency domains, and color space transformations. First system used rotational signatures based on a correlation operator;minimum distance was used for the classification task. Second system computed the seven Hu invariants from binary images;these descriptors fed to a Multi-Layer Perceptron (MLP) in order to recognize the 9 different classes. First system achieves 100% of recognition rate with leaving-one-out validation and second experiment performs 96.7% of recognition rate with Hu moments and 100% using 36 normalized moments and k-fold cross validation.展开更多
Deep Learning is a powerful technique that is widely applied to Image Recognition and Natural Language Processing tasks amongst many other tasks. In this work, we propose an efficient technique to utilize pre-trained ...Deep Learning is a powerful technique that is widely applied to Image Recognition and Natural Language Processing tasks amongst many other tasks. In this work, we propose an efficient technique to utilize pre-trained Convolutional Neural Network (CNN) architectures to extract powerful features from images for object recognition purposes. We have built on the existing concept of extending the learning from pre-trained CNNs to new databases through activations by proposing to consider multiple deep layers. We have exploited the progressive learning that happens at the various intermediate layers of the CNNs to construct Deep Multi-Layer (DM-L) based Feature Extraction vectors to achieve excellent object recognition performance. Two popular pre-trained CNN architecture models i.e. the VGG_16 and VGG_19 have been used in this work to extract the feature sets from 3 deep fully connected multiple layers namely “fc6”, “fc7” and “fc8” from inside the models for object recognition purposes. Using the Principal Component Analysis (PCA) technique, the Dimensionality of the DM-L feature vectors has been reduced to form powerful feature vectors that have been fed to an external Classifier Ensemble for classification instead of the Softmax based classification layers of the two original pre-trained CNN models. The proposed DM-L technique has been applied to the Benchmark Caltech-101 object recognition database. Conventional wisdom may suggest that feature extractions based on the deepest layer i.e. “fc8” compared to “fc6” will result in the best recognition performance but our results have proved it otherwise for the two considered models. Our experiments have revealed that for the two models under consideration, the “fc6” based feature vectors have achieved the best recognition performance. State-of-the-Art recognition performances of 91.17% and 91.35% have been achieved by utilizing the “fc6” based feature vectors for the VGG_16 and VGG_19 models respectively. The recognition performance has been achieved by considering 30 sample images per class whereas the proposed system is capable of achieving improved performance by considering all sample images per class. Our research shows that for feature extraction based on CNNs, multiple layers should be considered and then the best layer can be selected that maximizes the recognition performance.展开更多
现有基于深度学习的农作物病害识别方法对网络浅层、中层、深层特征中包含的判别信息挖掘不够,且提取的农作物病害图像显著性特征大多不足,为了更加有效地提取农作物病害图像中的判别特征,提高农作物病害识别精度,提出一种基于多层信息...现有基于深度学习的农作物病害识别方法对网络浅层、中层、深层特征中包含的判别信息挖掘不够,且提取的农作物病害图像显著性特征大多不足,为了更加有效地提取农作物病害图像中的判别特征,提高农作物病害识别精度,提出一种基于多层信息融合和显著性特征增强的农作物病害识别网络(Crop disease recognition network based on multi-layer information fusion and saliency feature enhancement,MISF-Net)。MISF-Net主要由ConvNext主干网络、多层信息融合模块、显著性特征增强模块组成。其中,ConvNext主干网络主要用于提取农作物病害图像的特征;多层信息融合模块主要用于提取和融合主干网络浅层、中层、深层特征中的判别信息;显著性特征增强模块主要用于增强农作物病害图像中的显著性判别特征。在农作物病害数据集AI challenger 2018及自制数据集RCP-Crops上的实验结果表明,MISF-Net的农作物病害识别准确率分别达到87.84%、95.41%,F1值分别达到87.72%、95.31%。展开更多
文摘This paper presents two systems for recognizing static signs (digits) from American Sign Language (ASL). These systems avoid the use color marks, or gloves, using instead, low-pass and high-pass filters in space and frequency domains, and color space transformations. First system used rotational signatures based on a correlation operator;minimum distance was used for the classification task. Second system computed the seven Hu invariants from binary images;these descriptors fed to a Multi-Layer Perceptron (MLP) in order to recognize the 9 different classes. First system achieves 100% of recognition rate with leaving-one-out validation and second experiment performs 96.7% of recognition rate with Hu moments and 100% using 36 normalized moments and k-fold cross validation.
文摘Deep Learning is a powerful technique that is widely applied to Image Recognition and Natural Language Processing tasks amongst many other tasks. In this work, we propose an efficient technique to utilize pre-trained Convolutional Neural Network (CNN) architectures to extract powerful features from images for object recognition purposes. We have built on the existing concept of extending the learning from pre-trained CNNs to new databases through activations by proposing to consider multiple deep layers. We have exploited the progressive learning that happens at the various intermediate layers of the CNNs to construct Deep Multi-Layer (DM-L) based Feature Extraction vectors to achieve excellent object recognition performance. Two popular pre-trained CNN architecture models i.e. the VGG_16 and VGG_19 have been used in this work to extract the feature sets from 3 deep fully connected multiple layers namely “fc6”, “fc7” and “fc8” from inside the models for object recognition purposes. Using the Principal Component Analysis (PCA) technique, the Dimensionality of the DM-L feature vectors has been reduced to form powerful feature vectors that have been fed to an external Classifier Ensemble for classification instead of the Softmax based classification layers of the two original pre-trained CNN models. The proposed DM-L technique has been applied to the Benchmark Caltech-101 object recognition database. Conventional wisdom may suggest that feature extractions based on the deepest layer i.e. “fc8” compared to “fc6” will result in the best recognition performance but our results have proved it otherwise for the two considered models. Our experiments have revealed that for the two models under consideration, the “fc6” based feature vectors have achieved the best recognition performance. State-of-the-Art recognition performances of 91.17% and 91.35% have been achieved by utilizing the “fc6” based feature vectors for the VGG_16 and VGG_19 models respectively. The recognition performance has been achieved by considering 30 sample images per class whereas the proposed system is capable of achieving improved performance by considering all sample images per class. Our research shows that for feature extraction based on CNNs, multiple layers should be considered and then the best layer can be selected that maximizes the recognition performance.
文摘现有基于深度学习的农作物病害识别方法对网络浅层、中层、深层特征中包含的判别信息挖掘不够,且提取的农作物病害图像显著性特征大多不足,为了更加有效地提取农作物病害图像中的判别特征,提高农作物病害识别精度,提出一种基于多层信息融合和显著性特征增强的农作物病害识别网络(Crop disease recognition network based on multi-layer information fusion and saliency feature enhancement,MISF-Net)。MISF-Net主要由ConvNext主干网络、多层信息融合模块、显著性特征增强模块组成。其中,ConvNext主干网络主要用于提取农作物病害图像的特征;多层信息融合模块主要用于提取和融合主干网络浅层、中层、深层特征中的判别信息;显著性特征增强模块主要用于增强农作物病害图像中的显著性判别特征。在农作物病害数据集AI challenger 2018及自制数据集RCP-Crops上的实验结果表明,MISF-Net的农作物病害识别准确率分别达到87.84%、95.41%,F1值分别达到87.72%、95.31%。