期刊文献+

基于分离式标签协同学习的YOLOv5多属性分类

YOLOv5 multi-attribute classification based on separable label collaborative learning
下载PDF
导出
摘要 针对图像分类任务中卷积网络提取图像细粒度特征能力不足、多属性之间的依赖关系无法识别的问题,提出一种基于YOLOv5的车辆多属性分类方法Multi-YOLOv5。该方法设计了多头非极大值抑制(Multi-NMS)和分离式标签损失(Separate-Loss)函数协同工作机制实现车辆的多属性分类任务,并采用卷积块注意力模块(CBAM)、SA(Shuffle Attention)和CoordConv方法重构了YOLOv5检测模型,分别从提升多属性特征能力提取、增强不同属性之间的关联关系、增强网络对位置信息的感知能力三方面提升模型对目标多属性分类的精准性。在VeRi等数据集上进行了训练与测试,实验结果表明,与基于GoogLeNet、残差网络(ResNet)、EfficientNet、ViT(Vision Transformer)等的网络结构相比,Multi-YOLOv5方法在目标的多属性分类方面取得了较好的识别结果,在VeRi数据集上,它的平均精度均值(mAP)达到了87.37%,较上述表现最佳的方法提高了4.47个百分点,且比原YOLOv5模型具有更好的鲁棒性,能为密集环境下的交通目标感知提供可靠的数据信息。 An Multi-YOLOv5 method was proposed for vehicle multi-attribute classification based on YOLOv5 to address the challenges of insufficient ability of convolutional networks to extract fine-grained features of images and inability to recognize dependencies between multiple attributes in image classification tasks.A collaborative working mechanism of Multi-head Non-Maximum Suppression(Multi-NMS)and separable label loss(Separate-Loss)function was designed to complete the multi-attribute classification task of vehicles.Additionally,the YOLOv5 detection model was reconstructed by using Convolutional Block Attention Module(CBAM),Shuffle Attention(SA),and CoordConv methods to enhance the ability of extracting multi-attribute features,strengthen the correlation between different attributes,and enhance the network’s perception of positional information,thereby improving the accuracy of the model in multi-attribute classification of objects.Finally,training and testing were conducted on datasets such as VeRi.Experimental results demonstrate that the Multi-YOLOv5 approach achieves superior recognition outcomes in multi-attribute classification of objects compared to network architectures including GoogLeNet,Residual Network(ResNet),EfficientNet,and Vision Transformer(ViT).The mean Average Precision(mAP)of Multi-YOLOv5 reaches 87.37%on VeRi dataset,with a remarkable improvement of 4.47 percentage points over the best-performing method mentioned above.Moreover,Multi-YOLOv5 exhibits better robustness compared to the original YOLOv5 model,thus providing reliable data information for traffic object perception in dense environments.
作者 李鑫 孟乔 皇甫俊逸 孟令辰 LI Xin;MENG Qiao;HUANGFU Junyi;MENG Lingchen(Department of Computer Technology and Applications,Qinghai University,Xining Qinghai 810016,China)
出处 《计算机应用》 CSCD 北大核心 2024年第5期1619-1628,共10页 journal of Computer Applications
基金 青海省自然科学基金资助项目(2023-ZJ-989Q)。
关键词 多属性分类 深度学习 多特征融合 注意力 YOLOv5 multi-attribute classification deep learning multi-feature fusion attention YOLOv5
  • 相关文献

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部