Breast cancer has become a killer of women's health nowadays.In order to exploit the potential representational capabilities of the models more comprehensively,we propose a multi-model fusion strategy.Specifically...Breast cancer has become a killer of women's health nowadays.In order to exploit the potential representational capabilities of the models more comprehensively,we propose a multi-model fusion strategy.Specifically,we combine two differently structured deep learning models,ResNet101 and Swin Transformer(SwinT),with the addition of the Convolutional Block Attention Module(CBAM)attention mechanism,which makes full use of SwinT's global context information modeling ability and ResNet101's local feature extraction ability,and additionally the cross entropy loss function is replaced by the focus loss function to solve the problem of unbalanced allocation of breast cancer data sets.The multi-classification recognition accuracies of the proposed fusion model under 40X,100X,200X and 400X BreakHis datasets are 97.50%,96.60%,96.30 and 96.10%,respectively.Compared with a single SwinT model and ResNet 101 model,the fusion model has higher accuracy and better generalization ability,which provides a more effective method for screening,diagnosis and pathological classification of female breast cancer.展开更多
As ocular computer-aided diagnostic(CAD)tools become more widely accessible,many researchers are developing deep learning(DL)methods to aid in ocular disease(OHD)diagnosis.Common eye diseases like cataracts(CATR),glau...As ocular computer-aided diagnostic(CAD)tools become more widely accessible,many researchers are developing deep learning(DL)methods to aid in ocular disease(OHD)diagnosis.Common eye diseases like cataracts(CATR),glaucoma(GLU),and age-related macular degeneration(AMD)are the focus of this study,which uses DL to examine their identification.Data imbalance and outliers are widespread in fundus images,which can make it difficult to apply manyDL algorithms to accomplish this analytical assignment.The creation of efficient and reliable DL algorithms is seen to be the key to further enhancing detection performance.Using the analysis of images of the color of the retinal fundus,this study offers a DL model that is combined with a one-of-a-kind concoction loss function(CLF)for the automated identification of OHD.This study presents a combination of focal loss(FL)and correntropy-induced loss functions(CILF)in the proposed DL model to improve the recognition performance of classifiers for biomedical data.This is done because of the good generalization and robustness of these two types of losses in addressing complex datasets with class imbalance and outliers.The classification performance of the DL model with our proposed loss function is compared to that of the baseline models using accuracy(ACU),recall(REC),specificity(SPF),Kappa,and area under the receiver operating characteristic curve(AUC)as the evaluation metrics.The testing shows that the method is reliable and efficient.展开更多
为了提高多视图深度估计结果精度,提出一种基于自适应空间特征增强的多视图深度估计算法。设计了由改进后的特征金字塔网络(feature pyramid network,FPN)和自适应空间特征增强(adaptive space feature enhancement,ASFE)组成的多尺度...为了提高多视图深度估计结果精度,提出一种基于自适应空间特征增强的多视图深度估计算法。设计了由改进后的特征金字塔网络(feature pyramid network,FPN)和自适应空间特征增强(adaptive space feature enhancement,ASFE)组成的多尺度特征提取模块,获取到具有全局上下文信息和位置信息的多尺度特征图像。通过残差学习网络对深度图进行优化,防止多次卷积操作出现重建边缘模糊的问题。通过分类的思想构建focal loss函数增强网络模型的判断能力。由实验结果可知,该算法在DTU(technical university of denmark)数据集上和CasMVSNet(Cascade MVSNet)算法相比,在整体精度误差、运行时间、显存资源占用上分别降低了14.08%、72.15%、4.62%。在Tanks and Temples数据集整体评价指标Mean上该模型优于其他算法,证明提出的基于自适应空间特征增强的多视图深度估计算法的有效性。展开更多
果园环境下柑橘的快速准确检测是自主采摘机器人作业的关键.针对现有的模型过于冗余、检测速度与精度不平衡等问题,提出一种轻量型果园环境果实检测方法.在YOLOv4算法的基础上引入焦点损失函数(Focal Loss)来提高模型在二分类检测任务...果园环境下柑橘的快速准确检测是自主采摘机器人作业的关键.针对现有的模型过于冗余、检测速度与精度不平衡等问题,提出一种轻量型果园环境果实检测方法.在YOLOv4算法的基础上引入焦点损失函数(Focal Loss)来提高模型在二分类检测任务中的负样本挖掘能力,并针对模型参数冗余等问题提出一种优化的模型剪枝方法.试验结果表明:提出的方法在果园环境中柑橘果实数据集检测得到的平均精度均值(mean average precision,M_(AP))达到94.22%,相较于YOLOv4模型提高了1.18%,模型参数减小了95.22%,模型尺寸为原来的4.84%,检测速度为原来的4.03倍.展开更多
基金By the National Natural Science Foundation of China(NSFC)(No.61772358),the National Key R&D Program Funded Project(No.2021YFE0105500),and the Jiangsu University‘Blue Project’.
文摘Breast cancer has become a killer of women's health nowadays.In order to exploit the potential representational capabilities of the models more comprehensively,we propose a multi-model fusion strategy.Specifically,we combine two differently structured deep learning models,ResNet101 and Swin Transformer(SwinT),with the addition of the Convolutional Block Attention Module(CBAM)attention mechanism,which makes full use of SwinT's global context information modeling ability and ResNet101's local feature extraction ability,and additionally the cross entropy loss function is replaced by the focus loss function to solve the problem of unbalanced allocation of breast cancer data sets.The multi-classification recognition accuracies of the proposed fusion model under 40X,100X,200X and 400X BreakHis datasets are 97.50%,96.60%,96.30 and 96.10%,respectively.Compared with a single SwinT model and ResNet 101 model,the fusion model has higher accuracy and better generalization ability,which provides a more effective method for screening,diagnosis and pathological classification of female breast cancer.
基金supported by the Deanship of Scientific Research,Vice Presidency forGraduate Studies and Scientific Research,King Faisal University,Saudi Arabia[Grant No.3,363].
文摘As ocular computer-aided diagnostic(CAD)tools become more widely accessible,many researchers are developing deep learning(DL)methods to aid in ocular disease(OHD)diagnosis.Common eye diseases like cataracts(CATR),glaucoma(GLU),and age-related macular degeneration(AMD)are the focus of this study,which uses DL to examine their identification.Data imbalance and outliers are widespread in fundus images,which can make it difficult to apply manyDL algorithms to accomplish this analytical assignment.The creation of efficient and reliable DL algorithms is seen to be the key to further enhancing detection performance.Using the analysis of images of the color of the retinal fundus,this study offers a DL model that is combined with a one-of-a-kind concoction loss function(CLF)for the automated identification of OHD.This study presents a combination of focal loss(FL)and correntropy-induced loss functions(CILF)in the proposed DL model to improve the recognition performance of classifiers for biomedical data.This is done because of the good generalization and robustness of these two types of losses in addressing complex datasets with class imbalance and outliers.The classification performance of the DL model with our proposed loss function is compared to that of the baseline models using accuracy(ACU),recall(REC),specificity(SPF),Kappa,and area under the receiver operating characteristic curve(AUC)as the evaluation metrics.The testing shows that the method is reliable and efficient.
文摘为了提高多视图深度估计结果精度,提出一种基于自适应空间特征增强的多视图深度估计算法。设计了由改进后的特征金字塔网络(feature pyramid network,FPN)和自适应空间特征增强(adaptive space feature enhancement,ASFE)组成的多尺度特征提取模块,获取到具有全局上下文信息和位置信息的多尺度特征图像。通过残差学习网络对深度图进行优化,防止多次卷积操作出现重建边缘模糊的问题。通过分类的思想构建focal loss函数增强网络模型的判断能力。由实验结果可知,该算法在DTU(technical university of denmark)数据集上和CasMVSNet(Cascade MVSNet)算法相比,在整体精度误差、运行时间、显存资源占用上分别降低了14.08%、72.15%、4.62%。在Tanks and Temples数据集整体评价指标Mean上该模型优于其他算法,证明提出的基于自适应空间特征增强的多视图深度估计算法的有效性。
文摘果园环境下柑橘的快速准确检测是自主采摘机器人作业的关键.针对现有的模型过于冗余、检测速度与精度不平衡等问题,提出一种轻量型果园环境果实检测方法.在YOLOv4算法的基础上引入焦点损失函数(Focal Loss)来提高模型在二分类检测任务中的负样本挖掘能力,并针对模型参数冗余等问题提出一种优化的模型剪枝方法.试验结果表明:提出的方法在果园环境中柑橘果实数据集检测得到的平均精度均值(mean average precision,M_(AP))达到94.22%,相较于YOLOv4模型提高了1.18%,模型参数减小了95.22%,模型尺寸为原来的4.84%,检测速度为原来的4.03倍.