期刊文献+
共找到9篇文章
< 1 >
每页显示 20 50 100
Traffic Sign Recognition for Autonomous Vehicle Using Optimized YOLOv7 and Convolutional Block Attention Module 被引量:1
1
作者 P.Kuppusamy M.Sanjay +1 位作者 P.V.Deepashree C.Iwendi 《Computers, Materials & Continua》 SCIE EI 2023年第10期445-466,共22页
The infrastructure and construction of roads are crucial for the economic and social development of a region,but traffic-related challenges like accidents and congestion persist.Artificial Intelligence(AI)and Machine ... The infrastructure and construction of roads are crucial for the economic and social development of a region,but traffic-related challenges like accidents and congestion persist.Artificial Intelligence(AI)and Machine Learning(ML)have been used in road infrastructure and construction,particularly with the Internet of Things(IoT)devices.Object detection in Computer Vision also plays a key role in improving road infrastructure and addressing trafficrelated problems.This study aims to use You Only Look Once version 7(YOLOv7),Convolutional Block Attention Module(CBAM),the most optimized object-detection algorithm,to detect and identify traffic signs,and analyze effective combinations of adaptive optimizers like Adaptive Moment estimation(Adam),Root Mean Squared Propagation(RMSprop)and Stochastic Gradient Descent(SGD)with the YOLOv7.Using a portion of German traffic signs for training,the study investigates the feasibility of adopting smaller datasets while maintaining high accuracy.The model proposed in this study not only improves traffic safety by detecting traffic signs but also has the potential to contribute to the rapid development of autonomous vehicle systems.The study results showed an impressive accuracy of 99.7%when using a batch size of 8 and the Adam optimizer.This high level of accuracy demonstrates the effectiveness of the proposed model for the image classification task of traffic sign recognition. 展开更多
关键词 Object detection traffic sign detection YOLOv7 convolutional block attention module road sign detection ADAM
下载PDF
ANC: Attention Network for COVID-19 Explainable Diagnosis Based on Convolutional Block Attention Module 被引量:9
2
作者 Yudong Zhang Xin Zhang Weiguo Zhu 《Computer Modeling in Engineering & Sciences》 SCIE EI 2021年第6期1037-1058,共22页
Aim: To diagnose COVID-19 more efficiently and more correctly, this study proposed a novel attention network forCOVID-19 (ANC). Methods: Two datasets were used in this study. An 18-way data augmentation was proposed t... Aim: To diagnose COVID-19 more efficiently and more correctly, this study proposed a novel attention network forCOVID-19 (ANC). Methods: Two datasets were used in this study. An 18-way data augmentation was proposed toavoid overfitting. Then, convolutional block attention module (CBAM) was integrated to our model, the structureof which is fine-tuned. Finally, Grad-CAM was used to provide an explainable diagnosis. Results: The accuracyof our ANC methods on two datasets are 96.32% ± 1.06%, and 96.00% ± 1.03%, respectively. Conclusions: Thisproposed ANC method is superior to 9 state-of-the-art approaches. 展开更多
关键词 Deep learning convolutional block attention module attention mechanism COVID-19 explainable diagnosis
下载PDF
MobileNet network optimization based on convolutional block attention module 被引量:3
3
作者 ZHAO Shuxu MEN Shiyao YUAN Lin 《Journal of Measurement Science and Instrumentation》 CAS CSCD 2022年第2期225-234,共10页
Deep learning technology is widely used in computer vision.Generally,a large amount of data is used to train the model weights in deep learning,so as to obtain a model with higher accuracy.However,massive data and com... Deep learning technology is widely used in computer vision.Generally,a large amount of data is used to train the model weights in deep learning,so as to obtain a model with higher accuracy.However,massive data and complex model structures require more calculating resources.Since people generally can only carry and use mobile and portable devices in application scenarios,neural networks have limitations in terms of calculating resources,size and power consumption.Therefore,the efficient lightweight model MobileNet is used as the basic network in this study for optimization.First,the accuracy of the MobileNet model is improved by adding methods such as the convolutional block attention module(CBAM)and expansion convolution.Then,the MobileNet model is compressed by using pruning and weight quantization algorithms based on weight size.Afterwards,methods such as Python crawlers and data augmentation are employed to create a garbage classification data set.Based on the above model optimization strategy,the garbage classification mobile terminal application is deployed on mobile phones and raspberry pies,realizing completing the garbage classification task more conveniently. 展开更多
关键词 MobileNet convolutional block attention module(CBAM) model pruning and quantization edge machine learning
下载PDF
Using Speaker-Specific Emotion Representations in Wav2vec 2.0-Based Modules for Speech Emotion Recognition
4
作者 Somin Park Mpabulungi Mark +1 位作者 Bogyung Park Hyunki Hong 《Computers, Materials & Continua》 SCIE EI 2023年第10期1009-1030,共22页
Speech emotion recognition is essential for frictionless human-machine interaction,where machines respond to human instructions with context-aware actions.The properties of individuals’voices vary with culture,langua... Speech emotion recognition is essential for frictionless human-machine interaction,where machines respond to human instructions with context-aware actions.The properties of individuals’voices vary with culture,language,gender,and personality.These variations in speaker-specific properties may hamper the performance of standard representations in downstream tasks such as speech emotion recognition(SER).This study demonstrates the significance of speaker-specific speech characteristics and how considering them can be leveraged to improve the performance of SER models.In the proposed approach,two wav2vec-based modules(a speaker-identification network and an emotion classification network)are trained with the Arcface loss.The speaker-identification network has a single attention block to encode an input audio waveform into a speaker-specific representation.The emotion classification network uses a wav2vec 2.0-backbone as well as four attention blocks to encode the same input audio waveform into an emotion representation.These two representations are then fused into a single vector representation containing emotion and speaker-specific information.Experimental results showed that the use of speaker-specific characteristics improves SER performance.Additionally,combining these with an angular marginal loss such as the Arcface loss improves intra-class compactness while increasing inter-class separability,as demonstrated by the plots of t-distributed stochastic neighbor embeddings(t-SNE).The proposed approach outperforms previous methods using similar training strategies,with a weighted accuracy(WA)of 72.14%and unweighted accuracy(UA)of 72.97%on the Interactive Emotional Dynamic Motion Capture(IEMOCAP)dataset.This demonstrates its effectiveness and potential to enhance human-machine interaction through more accurate emotion recognition in speech. 展开更多
关键词 attention block IEMOCAP dataset speaker-specific representation speech emotion recognition wav2vec 2.0
下载PDF
Lightweight Surface Litter Detection Algorithm Based on Improved YOLOv5s 被引量:1
5
作者 Zunliang Chen Chengxu Huang +1 位作者 Lucheng Duan Baohua Tan 《Computers, Materials & Continua》 SCIE EI 2023年第7期1085-1102,共18页
In response to the problem of the high cost and low efficiency of traditional water surface litter cleanup through manpower,a lightweight water surface litter detection algorithm based on improved YOLOv5s is proposed ... In response to the problem of the high cost and low efficiency of traditional water surface litter cleanup through manpower,a lightweight water surface litter detection algorithm based on improved YOLOv5s is proposed to provide core technical support for real-time water surface litter detection by water surface litter cleanup vessels.The method reduces network parameters by introducing the deep separable convolution GhostConv in the lightweight network GhostNet to substitute the ordinary convolution in the original YOLOv5s feature extraction and fusion network;introducing the C3Ghost module to substitute the C3 module in the original backbone and neck networks to further reduce computational effort.Using a Convolutional Block Attention Mechanism(CBAM)module in the backbone network to strengthen the network’s ability to extract significant target features from images.Finally,the loss function is optimized using the Focal-EIoU loss func-tion to improve the convergence speed and model accuracy.The experimental results illustrate that the improved algorithm outperforms the original Yolov5s in all aspects of the homemade water surface litter dataset and has certain advantages over some current mainstream algorithms in terms of model size,detection accuracy,and speed,which can deal with the problems of real-time detection of water surface litter in real life. 展开更多
关键词 Surface litter detection LIGHTWEIGHT YOLOv5s GhostNet deep separable convolution convolutional block attention mechanism(CBAM)
下载PDF
Classifying Hematoxylin and Eosin Images Using a Super-Resolution Segmentor and a Deep Ensemble Classifier
6
作者 P.Sabitha G.Meeragandhi 《Intelligent Automation & Soft Computing》 SCIE 2023年第8期1983-2000,共18页
Developing an automatic and credible diagnostic system to analyze the type,stage,and level of the liver cancer from Hematoxylin and Eosin(H&E)images is a very challenging and time-consuming endeavor,even for exper... Developing an automatic and credible diagnostic system to analyze the type,stage,and level of the liver cancer from Hematoxylin and Eosin(H&E)images is a very challenging and time-consuming endeavor,even for experienced pathologists,due to the non-uniform illumination and artifacts.Albeit several Machine Learning(ML)and Deep Learning(DL)approaches are employed to increase the performance of automatic liver cancer diagnostic systems,the classi-fication accuracy of these systems still needs significant improvement to satisfy the real-time requirement of the diagnostic situations.In this work,we present a new Ensemble Classifier(hereafter called ECNet)to classify the H&E stained liver histopathology images effectively.The proposed model employs a Dropout Extreme Learning Machine(DrpXLM)and the Enhanced Convolutional Block Attention Modules(ECBAM)based residual network.ECNet applies Voting Mechanism(VM)to integrate the decisions of individual classifiers using the average of probabilities rule.Initially,the nuclei regions in the H&E stain are seg-mented through Super-resolution Convolutional Networks(SrCN),and then these regions are fed into the ensemble DL network for classification.The effectiveness of the proposed model is carefully studied on real-world datasets.The results of our meticulous experiments on the Kasturba Medical College(KMC)liver dataset reveal that the proposed ECNet significantly outperforms other existing classifica-tion networks with better accuracy,sensitivity,specificity,precision,and Jaccard Similarity Score(JSS)of 96.5%,99.4%,89.7%,95.7%,and 95.2%,respectively.We obtain similar results from ECNet when applied to The Cancer Genome Atlas Liver Hepatocellular Carcinoma(TCGA-LIHC)dataset regarding accuracy(96.3%),sensitivity(97.5%),specificity(93.2%),precision(97.5%),and JSS(95.1%).More importantly,the proposed ECNet system consumes only 12.22 s for training and 1.24 s for testing.Also,we carry out the Wilcoxon statistical test to determine whether the ECNet provides a considerable improvement with respect to evaluation metrics or not.From extensive empirical analysis,we can conclude that our ECNet is the better liver cancer diagnostic model related to state-of-the-art classifiers. 展开更多
关键词 Convolutional block attention modules dropout ELM ensemble classifier liver cancer segmentation voting mechanism
下载PDF
基于注意力特征融合的SqueezeNet细粒度图像分类模型 被引量:7
7
作者 李明悦 何乐生 +1 位作者 雷晨 龚友梅 《云南大学学报(自然科学版)》 CAS CSCD 北大核心 2021年第5期868-876,共9页
针对现有细粒度图像分类算法普遍存在的模型结构复杂、参数多、分类准确率较低等问题,提出一种注意力特征融合的SqueezeNet细粒度图像分类模型.通过对现有细粒度图像分类算法和轻量级卷积神经网络的分析,首先使用3个典型的预训练轻量级... 针对现有细粒度图像分类算法普遍存在的模型结构复杂、参数多、分类准确率较低等问题,提出一种注意力特征融合的SqueezeNet细粒度图像分类模型.通过对现有细粒度图像分类算法和轻量级卷积神经网络的分析,首先使用3个典型的预训练轻量级卷积神经网络,对其微调后在公开的细粒度图像数据集上进行验证,经比较后选择了模型性能最佳的SqueezeNet作为图像的特征提取器;然后将两个具有注意力机制的卷积模块嵌入至SqueezeNet网络的每个Fire模块;接着提取出改进后的SqueezeNet的中间层特征进行双线性融合形成新的注意力特征图,与网络的全局特征再融合后分类;最后通过实验对比和可视化分析,网络嵌入Convolution Block Attention Module(CBAM)模块的分类准确率在鸟类、汽车、飞机数据集上依次提高了8.96%、4.89%和5.85%,嵌入Squeeze-and-Excitation(SE)模块的分类准确率依次提高了9.81%、4.52%和2.30%,且新模型在参数量、运行效率等方面比现有算法更具优势. 展开更多
关键词 细粒度图像分类 轻量级卷积神经网络 SqueezeNet 注意力机制 Convolution block attention Module(CBAM) Squeeze-and-Excitation(SE) 特征融合
下载PDF
Improved edge lightweight YOLOv4 and its application in on-site power system work 被引量:5
8
作者 Kexin Li Liang Qin +3 位作者 Qiang Li Feng Zhao Zhongping Xu Kaipei Liu 《Global Energy Interconnection》 EI CAS CSCD 2022年第2期168-180,共13页
A“cloud-edge-end”collaborative system architecture is adopted for real-time security management of power system on-site work,and mobile edge computing equipment utilizes lightweight intelligent recognition algorithm... A“cloud-edge-end”collaborative system architecture is adopted for real-time security management of power system on-site work,and mobile edge computing equipment utilizes lightweight intelligent recognition algorithms for on-site risk assessment and alert.Owing to its lightweight and fast speed,YOLOv4-Tiny is often deployed on edge computing equipment for real-time video stream detection;however,its accuracy is relatively low.This study proposes an improved YOLOv4-Tiny algorithm based on attention mechanism and optimized training methods,achieving higher accuracy without compromising the speed.Specifically,a convolution block attention module branch is added to the backbone network to enhance the feature extraction capability and an efficient channel attention mechanism is added in the neck network to improve feature utilization.Moreover,three optimized training methods:transfer learning,mosaic data augmentation,and label smoothing are used to improve the training effect of this improved algorithm.Finally,an edge computing equipment experimental platform equipped with an NVIDIA Jetson Xavier NX chip is established and the newly developed algorithm is tested on it.According to the results,the speed of the improved YOLOv4-Tiny algorithm in detecting on-site dress code compliance datasets is 17.25 FPS,and the mean average precision(mAP)is increased from 70.89%to 85.03%. 展开更多
关键词 On-site power system work YOLOv4-Tiny Convolution block attention mechanism Efficient channel attention Optimized training methods.
下载PDF
Foreground Segmentation Network with Enhanced Attention
9
作者 姜锐 朱瑞祥 +1 位作者 蔡萧萃 苏虎 《Journal of Shanghai Jiaotong university(Science)》 EI 2023年第3期360-369,共10页
Moving object segmentation (MOS) is one of the essential functions of the vision system of all robots,including medical robots. Deep learning-based MOS methods, especially deep end-to-end MOS methods, are actively inv... Moving object segmentation (MOS) is one of the essential functions of the vision system of all robots,including medical robots. Deep learning-based MOS methods, especially deep end-to-end MOS methods, are actively investigated in this field. Foreground segmentation networks (FgSegNets) are representative deep end-to-endMOS methods proposed recently. This study explores a new mechanism to improve the spatial feature learningcapability of FgSegNets with relatively few brought parameters. Specifically, we propose an enhanced attention(EA) module, a parallel connection of an attention module and a lightweight enhancement module, with sequentialattention and residual attention as special cases. We also propose integrating EA with FgSegNet_v2 by taking thelightweight convolutional block attention module as the attention module and plugging EA module after the twoMaxpooling layers of the encoder. The derived new model is named FgSegNet_v2 EA. The ablation study verifiesthe effectiveness of the proposed EA module and integration strategy. The results on the CDnet2014 dataset,which depicts human activities and vehicles captured in different scenes, show that FgSegNet_v2 EA outperformsFgSegNet_v2 by 0.08% and 14.5% under the settings of scene dependent evaluation and scene independent evaluation, respectively, which indicates the positive effect of EA on improving spatial feature learning capability ofFgSegNet_v2. 展开更多
关键词 human-computer interaction moving object segmentation foreground segmentation network enhanced attention convolutional block attention module
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部