期刊文献+
共找到23篇文章
< 1 2 >
每页显示 20 50 100
TSCND:Temporal Subsequence-Based Convolutional Network with Difference for Time Series Forecasting
1
作者 Haoran Huang Weiting Chen Zheming Fan 《Computers, Materials & Continua》 SCIE EI 2024年第3期3665-3681,共17页
Time series forecasting plays an important role in various fields, such as energy, finance, transport, and weather. Temporal convolutional networks (TCNs) based on dilated causal convolution have been widely used in t... Time series forecasting plays an important role in various fields, such as energy, finance, transport, and weather. Temporal convolutional networks (TCNs) based on dilated causal convolution have been widely used in time series forecasting. However, two problems weaken the performance of TCNs. One is that in dilated casual convolution, causal convolution leads to the receptive fields of outputs being concentrated in the earlier part of the input sequence, whereas the recent input information will be severely lost. The other is that the distribution shift problem in time series has not been adequately solved. To address the first problem, we propose a subsequence-based dilated convolution method (SDC). By using multiple convolutional filters to convolve elements of neighboring subsequences, the method extracts temporal features from a growing receptive field via a growing subsequence rather than a single element. Ultimately, the receptive field of each output element can cover the whole input sequence. To address the second problem, we propose a difference and compensation method (DCM). The method reduces the discrepancies between and within the input sequences by difference operations and then compensates the outputs for the information lost due to difference operations. Based on SDC and DCM, we further construct a temporal subsequence-based convolutional network with difference (TSCND) for time series forecasting. The experimental results show that TSCND can reduce prediction mean squared error by 7.3% and save runtime, compared with state-of-the-art models and vanilla TCN. 展开更多
关键词 DIFFERENCE data prediction time series temporal convolutional network dilated convolution
下载PDF
A Lightweight Convolutional Neural Network with Hierarchical Multi-Scale Feature Fusion for Image Classification
2
作者 Adama Dembele Ronald Waweru Mwangi Ananda Omutokoh Kube 《Journal of Computer and Communications》 2024年第2期173-200,共28页
Convolutional neural networks (CNNs) are widely used in image classification tasks, but their increasing model size and computation make them challenging to implement on embedded systems with constrained hardware reso... Convolutional neural networks (CNNs) are widely used in image classification tasks, but their increasing model size and computation make them challenging to implement on embedded systems with constrained hardware resources. To address this issue, the MobileNetV1 network was developed, which employs depthwise convolution to reduce network complexity. MobileNetV1 employs a stride of 2 in several convolutional layers to decrease the spatial resolution of feature maps, thereby lowering computational costs. However, this stride setting can lead to a loss of spatial information, particularly affecting the detection and representation of smaller objects or finer details in images. To maintain the trade-off between complexity and model performance, a lightweight convolutional neural network with hierarchical multi-scale feature fusion based on the MobileNetV1 network is proposed. The network consists of two main subnetworks. The first subnetwork uses a depthwise dilated separable convolution (DDSC) layer to learn imaging features with fewer parameters, which results in a lightweight and computationally inexpensive network. Furthermore, depthwise dilated convolution in DDSC layer effectively expands the field of view of filters, allowing them to incorporate a larger context. The second subnetwork is a hierarchical multi-scale feature fusion (HMFF) module that uses parallel multi-resolution branches architecture to process the input feature map in order to extract the multi-scale feature information of the input image. Experimental results on the CIFAR-10, Malaria, and KvasirV1 datasets demonstrate that the proposed method is efficient, reducing the network parameters and computational cost by 65.02% and 39.78%, respectively, while maintaining the network performance compared to the MobileNetV1 baseline. 展开更多
关键词 MobileNet Image Classification Lightweight convolutional Neural Network Depthwise Dilated Separable Convolution Hierarchical Multi-Scale Feature Fusion
下载PDF
Advanced Face Mask Detection Model Using Hybrid Dilation Convolution Based Method 被引量:1
3
作者 Shaohan Wang Xiangyu Wang Xin Guo 《Journal of Software Engineering and Applications》 2023年第1期1-19,共19页
A face-mask object detection model incorporating hybrid dilation convolutional network termed ResNet Hybrid-dilation-convolution Face-mask-detector (RHF) is proposed in this paper. Furthermore, a lightweight face-mask... A face-mask object detection model incorporating hybrid dilation convolutional network termed ResNet Hybrid-dilation-convolution Face-mask-detector (RHF) is proposed in this paper. Furthermore, a lightweight face-mask dataset named Light Masked Face Dataset (LMFD) and a medium-sized face-mask dataset named Masked Face Dataset (MFD) with data augmentation methods applied is also constructed in this paper. The hybrid dilation convolutional network is able to expand the perception of the convolutional kernel without concern about the discontinuity of image information during the convolution process. For the given two datasets being constructed above, the trained models are significantly optimized in terms of detection performance, training time, and other related metrics. By using the MFD dataset of 55,905 images, the RHF model requires roughly 10 hours less training time compared to ResNet50 with better detection results with mAP of 93.45%. 展开更多
关键词 Face Mask Detection Object Detection Hybrid dilation Convolution Computer Vision
下载PDF
Convolution-Transformer for Image Feature Extraction
4
作者 Lirong Yin Lei Wang +10 位作者 Siyu Lu Ruiyang Wang Youshuai Yang Bo Yang Shan Liu Ahmed AlSanad Salman A.AlQahtani Zhengtong Yin Xiaolu Li Xiaobing Chen Wenfeng Zheng 《Computer Modeling in Engineering & Sciences》 SCIE EI 2024年第10期87-106,共20页
This study addresses the limitations of Transformer models in image feature extraction,particularly their lack of inductive bias for visual structures.Compared to Convolutional Neural Networks(CNNs),the Transformers a... This study addresses the limitations of Transformer models in image feature extraction,particularly their lack of inductive bias for visual structures.Compared to Convolutional Neural Networks(CNNs),the Transformers are more sensitive to different hyperparameters of optimizers,which leads to a lack of stability and slow convergence.To tackle these challenges,we propose the Convolution-based Efficient Transformer Image Feature Extraction Network(CEFormer)as an enhancement of the Transformer architecture.Our model incorporates E-Attention,depthwise separable convolution,and dilated convolution to introduce crucial inductive biases,such as translation invariance,locality,and scale invariance,into the Transformer framework.Additionally,we implement a lightweight convolution module to process the input images,resulting in faster convergence and improved stability.This results in an efficient convolution combined Transformer image feature extraction network.Experimental results on the ImageNet1k Top-1 dataset demonstrate that the proposed network achieves better accuracy while maintaining high computational speed.It achieves up to 85.0%accuracy across various model sizes on image classification,outperforming various baseline models.When integrated into the Mask Region-ConvolutionalNeuralNetwork(R-CNN)framework as a backbone network,CEFormer outperforms other models and achieves the highest mean Average Precision(mAP)scores.This research presents a significant advancement in Transformer-based image feature extraction,balancing performance and computational efficiency. 展开更多
关键词 TRANSFORMER E-Attention depth convolution dilated convolution CEFormer
下载PDF
DcNet: Dilated Convolutional Neural Networks for Side-Scan Sonar Image Semantic Segmentation 被引量:2
5
作者 ZHAO Xiaohong QIN Rixia +3 位作者 ZHANG Qilei YU Fei WANG Qi HE Bo 《Journal of Ocean University of China》 SCIE CAS CSCD 2021年第5期1089-1096,共8页
In ocean explorations,side-scan sonar(SSS)plays a very important role and can quickly depict seabed topography.As-sembling the SSS to an autonomous underwater vehicle(AUV)and performing semantic segmentation of an SSS... In ocean explorations,side-scan sonar(SSS)plays a very important role and can quickly depict seabed topography.As-sembling the SSS to an autonomous underwater vehicle(AUV)and performing semantic segmentation of an SSS image in real time can realize online submarine geomorphology or target recognition,which is conducive to submarine detection.However,because of the complexity of the marine environment,various noises in the ocean pollute the sonar image,which also encounters the intensity inhomogeneity problem.In this paper,we propose a novel neural network architecture named dilated convolutional neural network(DcNet)that can run in real time while addressing the above-mentioned issues and providing accurate semantic segmentation.The proposed architecture presents an encoder-decoder network to gradually reduce the spatial dimension of the input image and recover the details of the target,respectively.The core of our network is a novel block connection named DCblock,which mainly uses dilated convolution and depthwise separable convolution between the encoder and decoder to attain more context while still retaining high accuracy.Furthermore,our proposed method performs a super-resolution reconstruction to enlarge the dataset with high-quality im-ages.We compared our network to other common semantic segmentation networks performed on an NVIDIA Jetson TX2 using our sonar image datasets.Experimental results show that while the inference speed of the proposed network significantly outperforms state-of-the-art architectures,the accuracy of our method is still comparable,which indicates its potential applications not only in AUVs equipped with SSS but also in marine exploration. 展开更多
关键词 side-scan sonar(SSS) semantic segmentation dilated convolutions SUPER-RESOLUTION
下载PDF
Multi⁃Scale Dilated Convolutional Neural Network for Hyperspectral Image Classification
6
作者 Shanshan Zheng Wen Liu +3 位作者 Rui Shan Jingyi Zhao Guoqian Jiang Zhi Zhang 《Journal of Harbin Institute of Technology(New Series)》 CAS 2021年第4期25-32,共8页
Aiming at the problem of image information loss,dilated convolution is introduced and a novel multi⁃scale dilated convolutional neural network(MDCNN)is proposed.Dilated convolution can polymerize image multi⁃scale inf... Aiming at the problem of image information loss,dilated convolution is introduced and a novel multi⁃scale dilated convolutional neural network(MDCNN)is proposed.Dilated convolution can polymerize image multi⁃scale information without reducing the resolution.The first layer of the network used spectral convolutional step to reduce dimensionality.Then the multi⁃scale aggregation extracted multi⁃scale features through applying dilated convolution and shortcut connection.The extracted features which represent properties of data were fed through Softmax to predict the samples.MDCNN achieved the overall accuracy of 99.58% and 99.92% on two public datasets,Indian Pines and Pavia University.Compared with four other existing models,the results illustrate that MDCNN can extract better discriminative features and achieve higher classification performance. 展开更多
关键词 multi⁃scale aggregation dilated convolution hyperspectral image classification(HSIC) shortcut connection
下载PDF
A Model for Helmet-Wearing Detection of Non-Motor Drivers Based on YOLOv5s 被引量:1
7
作者 Hongyu Lin Feng Jiang +3 位作者 Yu Jiang Huiyin Luo Jian Yao Jiaxin Liu 《Computers, Materials & Continua》 SCIE EI 2023年第6期5321-5336,共16页
Detecting non-motor drivers’helmets has significant implications for traffic control.Currently,most helmet detection methods are susceptible to the complex background and need more accuracy and better robustness of s... Detecting non-motor drivers’helmets has significant implications for traffic control.Currently,most helmet detection methods are susceptible to the complex background and need more accuracy and better robustness of small object detection,which are unsuitable for practical application scenar-ios.Therefore,this paper proposes a new helmet-wearing detection algorithm based on the You Only Look Once version 5(YOLOv5).First,the Dilated convolution In Coordinate Attention(DICA)layer is added to the backbone network.DICA combines the coordinated attention mechanism with atrous convolution to replace the original convolution layer,which can increase the perceptual field of the network to get more contextual information.Also,it can reduce the network’s learning of unnecessary features in the background and get attention to small objects.Second,the Rebuild Bidirectional Feature Pyramid Network(Re-BiFPN)is used as a feature extraction network.Re-BiFPN uses cross-scale feature fusion to combine the semantic information features at the high level with the spatial information features at the bottom level,which facilitates the model to learn object features at different scales.Verified on the proposed“Helmet Wearing dataset for Non-motor Drivers(HWND),”the results show that the proposed model is superior to the current detection algorithms,with the mean average precision(mAP)of 94.3%under complex background. 展开更多
关键词 Helmet-wearing detection dilated convolution feature pyramid network feature fusion
下载PDF
An improved deep dilated convolutional neural network for seismic facies interpretation
8
作者 Na-Xia Yang Guo-Fa Li +2 位作者 Ting-Hui Li Dong-Feng Zhao Wei-Wei Gu 《Petroleum Science》 SCIE EI CAS 2024年第3期1569-1583,共15页
With the successful application and breakthrough of deep learning technology in image segmentation,there has been continuous development in the field of seismic facies interpretation using convolutional neural network... With the successful application and breakthrough of deep learning technology in image segmentation,there has been continuous development in the field of seismic facies interpretation using convolutional neural networks.These intelligent and automated methods significantly reduce manual labor,particularly in the laborious task of manually labeling seismic facies.However,the extensive demand for training data imposes limitations on their wider application.To overcome this challenge,we adopt the UNet architecture as the foundational network structure for seismic facies classification,which has demonstrated effective segmentation results even with small-sample training data.Additionally,we integrate spatial pyramid pooling and dilated convolution modules into the network architecture to enhance the perception of spatial information across a broader range.The seismic facies classification test on the public data from the F3 block verifies the superior performance of our proposed improved network structure in delineating seismic facies boundaries.Comparative analysis against the traditional UNet model reveals that our method achieves more accurate predictive classification results,as evidenced by various evaluation metrics for image segmentation.Obviously,the classification accuracy reaches an impressive 96%.Furthermore,the results of seismic facies classification in the seismic slice dimension provide further confirmation of the superior performance of our proposed method,which accurately defines the range of different seismic facies.This approach holds significant potential for analyzing geological patterns and extracting valuable depositional information. 展开更多
关键词 Seismic facies interpretation Dilated convolution Spatial pyramid pooling Internal feature maps Compound loss function
下载PDF
1D-CNN:Speech Emotion Recognition System Using a Stacked Network with Dilated CNN Features 被引量:5
9
作者 Mustaqeem Soonil Kwon 《Computers, Materials & Continua》 SCIE EI 2021年第6期4039-4059,共21页
Emotion recognition from speech data is an active and emerging area of research that plays an important role in numerous applications,such as robotics,virtual reality,behavior assessments,and emergency call centers.Re... Emotion recognition from speech data is an active and emerging area of research that plays an important role in numerous applications,such as robotics,virtual reality,behavior assessments,and emergency call centers.Recently,researchers have developed many techniques in this field in order to ensure an improvement in the accuracy by utilizing several deep learning approaches,but the recognition rate is still not convincing.Our main aim is to develop a new technique that increases the recognition rate with reasonable cost computations.In this paper,we suggested a new technique,which is a one-dimensional dilated convolutional neural network(1D-DCNN)for speech emotion recognition(SER)that utilizes the hierarchical features learning blocks(HFLBs)with a bi-directional gated recurrent unit(BiGRU).We designed a one-dimensional CNN network to enhance the speech signals,which uses a spectral analysis,and to extract the hidden patterns from the speech signals that are fed into a stacked one-dimensional dilated network that are called HFLBs.Each HFLB contains one dilated convolution layer(DCL),one batch normalization(BN),and one leaky_relu(Relu)layer in order to extract the emotional features using a hieratical correlation strategy.Furthermore,the learned emotional features are feed into a BiGRU in order to adjust the global weights and to recognize the temporal cues.The final state of the deep BiGRU is passed from a softmax classifier in order to produce the probabilities of the emotions.The proposed model was evaluated over three benchmarked datasets that included the IEMOCAP,EMO-DB,and RAVDESS,which achieved 72.75%,91.14%,and 78.01%accuracy,respectively. 展开更多
关键词 Affective computing one-dimensional dilated convolutional neural network emotion recognition gated recurrent unit raw audio clips
下载PDF
Long Text Classification Algorithm Using a Hybrid Model of Bidirectional Encoder Representation from Transformers-Hierarchical Attention Networks-Dilated Convolutions Network 被引量:1
10
作者 赵媛媛 高世宁 +1 位作者 刘洋 宫晓蕙 《Journal of Donghua University(English Edition)》 CAS 2021年第4期341-350,共10页
Text format information is full of most of the resources of Internet,which puts forward higher and higher requirements for the accuracy of text classification.Therefore,in this manuscript,firstly,we design a hybrid mo... Text format information is full of most of the resources of Internet,which puts forward higher and higher requirements for the accuracy of text classification.Therefore,in this manuscript,firstly,we design a hybrid model of bidirectional encoder representation from transformers-hierarchical attention networks-dilated convolutions networks(BERT_HAN_DCN)which based on BERT pre-trained model with superior ability of extracting characteristic.The advantages of HAN model and DCN model are taken into account which can help gain abundant semantic information,fusing context semantic features and hierarchical characteristics.Secondly,the traditional softmax algorithm increases the learning difficulty of the same kind of samples,making it more difficult to distinguish similar features.Based on this,AM-softmax is introduced to replace the traditional softmax.Finally,the fused model is validated,which shows superior performance in the accuracy rate and F1-score of this hybrid model on two datasets and the experimental analysis shows the general single models such as HAN,DCN,based on BERT pre-trained model.Besides,the improved AM-softmax network model is superior to the general softmax network model. 展开更多
关键词 long text classification dilated convolution BERT fusing context semantic features hierarchical characteristics BERT_HAN_DCN AM-softmax
下载PDF
A Hybrid CNN for Image Denoising 被引量:3
11
作者 Menghua Zheng Keyan Zhi +2 位作者 Jiawen Zeng Chunwei Tian Lei You 《Journal of Artificial Intelligence and Technology》 2022年第3期93-99,共7页
Deep convolutional neural networks(CNNs)with strong learning abilities have been used in the field of image denoising.However,some CNNs depend on a single deep network to train an image denoising model,which will have... Deep convolutional neural networks(CNNs)with strong learning abilities have been used in the field of image denoising.However,some CNNs depend on a single deep network to train an image denoising model,which will have poor performance in complex screens.To address this problem,we propose a hybrid denoising CNN(HDCNN).HDCNN is composed of a dilated block(DB),RepVGG block(RVB),feature refinement block(FB),and a single convolution.DB combines a dilated convolution,batch normalization(BN),common convolutions,and activation function of ReLU to obtain more context information.RVB uses parallel combination of convolution,BN,and ReLU to extract complementary width features.FB is used to obtain more accurate information via refining obtained feature from the RVB.A single convolution collaborates a residual learning operation to construct a clean image.These key components make the HDCNN have good performance in image denoising.Experiment shows that the proposed HDCNN enjoys good denoising effect in public data sets. 展开更多
关键词 CNN dilated convolutions image denoising RepVGG
下载PDF
An image compressed sensing algorithm based on adaptive nonlinear network
12
作者 郭媛 陈炜 敬世伟 《Chinese Physics B》 SCIE EI CAS CSCD 2020年第5期243-253,共11页
Traditional compressed sensing algorithm is used to reconstruct images by iteratively optimizing a small number of measured values.The computation is complex and the reconstruction time is long.The deep learning-based... Traditional compressed sensing algorithm is used to reconstruct images by iteratively optimizing a small number of measured values.The computation is complex and the reconstruction time is long.The deep learning-based compressed sensing algorithm can greatly shorten the reconstruction time,but the algorithm emphasis is placed on reconstructing the network part mostly.The random measurement matrix cannot measure the image features well,which leads the reconstructed image quality to be improved limitedly.Two kinds of networks are proposed for solving this problem.The first one is ReconNet’s improved network IReconNet,which replaces the traditional linear random measurement matrix with an adaptive nonlinear measurement network.The reconstruction quality and anti-noise performance are greatly improved.Because the measured values extracted by the measurement network also retain the characteristics of image spatial information,the image is reconstructed by bilinear interpolation algorithm(Bilinear)and dilate convolution.Therefore a second network USDCNN is proposed.On the BSD500 dataset,the sampling rates are 0.25,0.10,0.04,and 0.01,the average peak signal-noise ratio(PSNR)of USDCNN is 1.62 dB,1.31 dB,1.47 dB,and 1.95 dB higher than that of MSRNet.Experiments show the average reconstruction time of USDCNN is 0.2705 s,0.3671 s,0.3602 s,and 0.3929 s faster than that of ReconNet.Moreover,there is also a great advantage in anti-noise performance. 展开更多
关键词 compressed sensing deep learning bilinear interpolation dilate convolution
下载PDF
A Deep Learning Approach for Crowd Counting in Highly Congested Scene
13
作者 Akbar Khan Kushsairy Abdul Kadir +5 位作者 Jawad Ali Shah Waleed Albattah Muhammad Saeed Haidawati Nasir Megat Norulazmi Megat Mohamed Noor Muhammad Haris Kaka Khel 《Computers, Materials & Continua》 SCIE EI 2022年第12期5825-5844,共20页
With the rapid progress of deep convolutional neural networks,several applications of crowd counting have been proposed and explored in the literature.In congested scene monitoring,a variety of crowd density estimatin... With the rapid progress of deep convolutional neural networks,several applications of crowd counting have been proposed and explored in the literature.In congested scene monitoring,a variety of crowd density estimating approaches has been developed.The understanding of highly congested scenes for crowd counting during Muslim gatherings of Hajj and Umrah is a challenging task,as a large number of individuals stand nearby and,it is hard for detection techniques to recognize them,as the crowd can vary from low density to high density.To deal with such highly congested scenes,we have proposed the Congested Scene Crowd Counting Network(CSCC-Net)using VGG-16 as a core network with its first ten layers due to its strong and robust transfer learning rate.A hole dilated convolutional neural network is used at the back end to widen the relevant field to extract a large range of information from the image without losing its original resolution.The dilated convolution neural network is mainly chosen to expand the kernel size without changing other parameters.Moreover,several loss functions have been applied to strengthen the evaluation accuracy of the model.Finally,the entire experiments have been evaluated using prominent data sets namely,ShanghaiTech parts A,B,UCF_CC_50,and UCF_QNRF.Our model has achieved remarkable results i.e.,68.0 and 9.0 MAE on ShanghaiTech parts A,B,199.1 MAE on UCF_CC_50,and 99.8 on UCF_QNRF data sets respectively. 展开更多
关键词 Deep learning congested scene crowd counting fully convolutional neural network dilated convolution
下载PDF
Application on Anomaly Detection of Geoelectric Field Based on Deep Learning
14
作者 WEI Lei AN Zhanghui +3 位作者 FAN Yingying CHEN Quan YUAN Lihua HOU Zeyu 《Earthquake Research in China》 CSCD 2020年第3期358-377,共20页
The deep learning method has made nurnerials achievements regarding anomaly detection in the field of time series.We introduce the speech production model in the field of artificial intelligence,changing the convoluti... The deep learning method has made nurnerials achievements regarding anomaly detection in the field of time series.We introduce the speech production model in the field of artificial intelligence,changing the convolution layer of the general convolution neural network to the residual element structure by adding identity mapping,and expanding the receptive domain of the model by using the dilated causal convolution.Based on the dilated causal convolution network and the method of log probability density function,the anomalous events are detected according to the anomaly scores.The validity of the method is verified by the simulation data,which is applied to the actual observed data on the observation staion of Pingliang geoeletric field in Gansu Province.The results show that one month before the Wenchuan M_S8.0,Lushan M_S7.0 and Minxian-Zhangxian M_S6.6 earthquakes,the daily cumulative error of log probability density of the predicted results in Pingliang Station suddenly decreases,which is consistent with the actual earthquake anomalies in a certain time range.After analyzing the combined factors including the spatial electromagnetic environment and the variation of micro fissures before the earthquake,we explain the possible causes of the anomalies in the geoelectric field of before the earthquake.The successful application of deep learning in observed data of the geoelectric field may behefit for improving the ultilization rate both the data and the efficiency of detection.Besides,it may provide technical support for more seismic research. 展开更多
关键词 Deep learning Time series Dilated causal convolution Geoelectric field Abnormal detection
下载PDF
Hard-rock tunnel lithology identification using multiscale dilated convolutional attention network based on tunnel face images
15
作者 Wenjun ZHANG Wuqi ZHANG +5 位作者 Gaole ZHANG Jun HUANG Minggeng LI Xiaohui WANG Fei YE Xiaoming GUAN 《Frontiers of Structural and Civil Engineering》 SCIE EI CSCD 2023年第12期1796-1812,共17页
For real-time classification of rock-masses in hard-rock tunnels,quick determination of the rock lithology on the tunnel face during construction is essential.Motivated by current breakthroughs in artificial intellige... For real-time classification of rock-masses in hard-rock tunnels,quick determination of the rock lithology on the tunnel face during construction is essential.Motivated by current breakthroughs in artificial intelligence technology in machine vision,a new automatic detection approach for classifying tunnel lithology based on tunnel face images was developed.The method benefits from residual learning for training a deep convolutional neural network(DCNN),and a multi-scale dilated convolutional attention block is proposed.The block with different dilation rates can provide various receptive fields,and thus it can extract multi-scale features.Moreover,the attention mechanism is utilized to select the salient features adaptively and further improve the performance of the model.In this study,an initial image data set made up of photographs of tunnel faces consisting of basalt,granite,siltstone,and tuff was first collected.After classifying and enhancing the training,validation,and testing data sets,a new image data set was generated.A comparison of the experimental findings demonstrated that the suggested approach outperforms previous classifiers in terms of various indicators,including accuracy,precision,recall,F1-score,and computing time.Finally,a visualization analysis was performed to explain the process of the network in the classification of tunnel lithology through feature extraction.Overall,this study demonstrates the potential of using artificial intelligence methods for in situ rock lithology classification utilizing geological images of the tunnel face. 展开更多
关键词 hard-rock tunnel face intelligent lithology identification multi-scale dilated convolutional attention network image classification deep learning
原文传递
Two Stages Segmentation Algorithm of Breast Tumor in DCE-MRI Based on Multi-Scale Feature and Boundary Attention Mechanism
16
作者 Bing Li Liangyu Wang +3 位作者 Xia Liu Hongbin Fan Bo Wang Shoudi Tong 《Computers, Materials & Continua》 SCIE EI 2024年第7期1543-1561,共19页
Nuclearmagnetic resonance imaging of breasts often presents complex backgrounds.Breast tumors exhibit varying sizes,uneven intensity,and indistinct boundaries.These characteristics can lead to challenges such as low a... Nuclearmagnetic resonance imaging of breasts often presents complex backgrounds.Breast tumors exhibit varying sizes,uneven intensity,and indistinct boundaries.These characteristics can lead to challenges such as low accuracy and incorrect segmentation during tumor segmentation.Thus,we propose a two-stage breast tumor segmentation method leveraging multi-scale features and boundary attention mechanisms.Initially,the breast region of interest is extracted to isolate the breast area from surrounding tissues and organs.Subsequently,we devise a fusion network incorporatingmulti-scale features and boundary attentionmechanisms for breast tumor segmentation.We incorporate multi-scale parallel dilated convolution modules into the network,enhancing its capability to segment tumors of various sizes through multi-scale convolution and novel fusion techniques.Additionally,attention and boundary detection modules are included to augment the network’s capacity to locate tumors by capturing nonlocal dependencies in both spatial and channel domains.Furthermore,a hybrid loss function with boundary weight is employed to address sample class imbalance issues and enhance the network’s boundary maintenance capability through additional loss.Themethod was evaluated using breast data from 207 patients at RuijinHospital,resulting in a 6.64%increase in Dice similarity coefficient compared to the benchmarkU-Net.Experimental results demonstrate the superiority of the method over other segmentation techniques,with fewer model parameters. 展开更多
关键词 Dynamic contrast-enhanced magnetic resonance imaging(DCE-MRI) breast tumor segmentation multi-scale dilated convolution boundary attention the hybrid loss function with boundary weight
下载PDF
KGTLIR:An Air Target Intention Recognition Model Based on Knowledge Graph and Deep Learning
17
作者 Bo Cao Qinghua Xing +2 位作者 Longyue Li Huaixi Xing Zhanfu Song 《Computers, Materials & Continua》 SCIE EI 2024年第7期1251-1275,共25页
As a core part of battlefield situational awareness,air target intention recognition plays an important role in modern air operations.Aiming at the problems of insufficient feature extraction and misclassification in ... As a core part of battlefield situational awareness,air target intention recognition plays an important role in modern air operations.Aiming at the problems of insufficient feature extraction and misclassification in intention recognition,this paper designs an air target intention recognition method(KGTLIR)based on Knowledge Graph and Deep Learning.Firstly,the intention recognition model based on Deep Learning is constructed to mine the temporal relationship of intention features using dilated causal convolution and the spatial relationship of intention features using a graph attention mechanism.Meanwhile,the accuracy,recall,and F1-score after iteration are introduced to dynamically adjust the sample weights to reduce the probability of misclassification.After that,an intention recognition model based on Knowledge Graph is constructed to predict the probability of the occurrence of different intentions of the target.Finally,the results of the two models are fused by evidence theory to obtain the target’s operational intention.Experiments show that the intention recognition accuracy of the KGTLIRmodel can reach 98.48%,which is not only better than most of the air target intention recognition methods,but also demonstrates better interpretability and trustworthiness. 展开更多
关键词 Dilated causal convolution graph attention mechanism intention recognition air targets knowledge graph
下载PDF
DTCC:Multi-level dilated convolution with transformer for weakly-supervised crowd counting
18
作者 Zhuangzhuang Miao Yong Zhang +2 位作者 Yuan Peng Haocheng Peng Baocai Yin 《Computational Visual Media》 SCIE EI CSCD 2023年第4期859-873,共15页
Crowd counting provides an important foundation for public security and urban management.Due to the existence of small targets and large density variations in crowd images,crowd counting is a challenging task.Mainstre... Crowd counting provides an important foundation for public security and urban management.Due to the existence of small targets and large density variations in crowd images,crowd counting is a challenging task.Mainstream methods usually apply convolution neural networks(CNNs)to regress a density map,which requires annotations of individual persons and counts.Weakly-supervised methods can avoid detailed labeling and only require counts as annotations of images,but existing methods fail to achieve satisfactory performance because a global perspective field and multi-level information are usually ignored.We propose a weakly-supervised method,DTCC,which effectively combines multi-level dilated convolution and transformer methods to realize end-to-end crowd counting.Its main components include a recursive swin transformer and a multi-level dilated convolution regression head.The recursive swin transformer combines a pyramid visual transformer with a fine-tuned recursive pyramid structure to capture deep multi-level crowd features,including global features.The multi-level dilated convolution regression head includes multi-level dilated convolution and a linear regression head for the feature extraction module.This module can capture both low-and high-level features simultaneously to enhance the receptive field.In addition,two regression head fusion mechanisms realize dynamic and mean fusion counting.Experiments on four well-known benchmark crowd counting datasets(UCF_CC_50,ShanghaiTech,UCF_QNRF,and JHU-Crowd++)show that DTCC achieves results superior to other weakly-supervised methods and comparable to fully-supervised methods. 展开更多
关键词 crowd counting TRANSFORMER dilated convolution global perspective field PYRAMID
原文传递
Automatic detection of sow estrus using a lightweight real-time detector and thermal images
19
作者 Haibo Zheng Hang Zhang +2 位作者 Shuang Song Yue Wang Tonghai Liu 《International Journal of Agricultural and Biological Engineering》 SCIE 2023年第3期194-207,共14页
Determination of ovulation time is one of the most important tasks in sow reproduction management.Temperature variation in the vulva of the sows can be used as a predictor of ovulation time.However,the skin temperatur... Determination of ovulation time is one of the most important tasks in sow reproduction management.Temperature variation in the vulva of the sows can be used as a predictor of ovulation time.However,the skin temperatures of sows in existing studies are obtained manually from infrared thermal images,posing an obstacle to the automatic prediction of ovulation time.In this study,an improved YOLO-V5s detector based on feature fusion and dilated convolution(FDYOLOV5s)was proposed for the automatic extraction of the vulva temperature of sows based on infrared thermal images.For the purpose of reducing the model complexity,the depthwise separable convolution and the modified lightweight ShuffleNet-V2 module were introduced in the backbone.Meanwhile,the feature fusion network structure of the model was simplified for efficiency,and a mixed dilated convolutional module was designed to obtain global features.The experimental results show that FD-YOLOV5s outperformed the other nine methods,with a mean average precision(mAP)of 99.1%,an average frame rate of 156.25 fps,and a model size of only 3.86 MB,indicating that the method effectively simplifies the model while ensuring detection accuracy.Using a linear regression between manual extraction and the results extracted using this method in randomly selected thermal images,the coefficients of determination for maximum and average vulvar temperatures reached 99.5%and 99.3%,respectively.The continuous vulva temperature of sows was obtained by the target detection algorithm,and the sow estrus detection was performed by the temperature trend and compared with the manually detected estrus results.The results showed that the sensitivity,specificity,and error rate of the estrus detection algorithm were 89.3%,94.5%,and 5.8%,respectively.The method achieves real-time and accurate extraction of sow vulva temperature and can be used for the automatic detection of sow estrus,which could be helpful for the automatic prediction of ovulation time. 展开更多
关键词 automatic estrus detection thermal images real-time detector vulva temperature mixed dilated convolutional
原文传递
Automatic Detection of Weld Defects in Pressure Vessel X-Ray Image Based on CNN
20
作者 XIAO Wenkai FENG Xiang +1 位作者 NAN Shuiyu ZHANG Linlin 《Wuhan University Journal of Natural Sciences》 CAS CSCD 2022年第6期489-498,共10页
The visual automatic detection method based on artificial intelligence has attracted more and more attention. In order to improve the performance of weld nondestructive defect detection,we propose DRepDet(Dilated RepP... The visual automatic detection method based on artificial intelligence has attracted more and more attention. In order to improve the performance of weld nondestructive defect detection,we propose DRepDet(Dilated RepPoints Detector). First, we analyze the weld defect dataset in detail and summarize the distribution characteristics of weld defect data, that is, the defect scale is very different and the aspect ratio distribution range is large. Second, according to the distribution characteristics of defect data, we design DResBlock module, and introduce dilated convolution with different dilated rates in the process of feature extraction to expand the receptive field and improve the detection performance of large-scale defects. Based on DResBlock and anchor-free detection framework RepPoints, we design DRepDet. Extensive experiments show that our proposed detector can detect 7 types of defects. When using combined dilated rate convolution network in detection, the AP50 and Recall50 of big defects are improved by 3.1% and 3.3% respectively, while the performance of small defects is not affected, almost the same or slightly improved. The final performance of the whole network is improved a large margin,with 6% AP50 and 4.2% Recall50 compared with Cascade RCNN and 1.4% AP50 and 2.9% Recall50 compared with RepPoints. 展开更多
关键词 nondestructive testing depth learning weld defect detection convolutional neural networks dilated convolution
原文传递
上一页 1 2 下一页 到第
使用帮助 返回顶部