期刊文献+
共找到8篇文章
< 1 >
每页显示 20 50 100
Unsupervised multi-modal image translation based on the squeeze-and-excitation mechanism and feature attention module
1
作者 胡振涛 HU Chonghao +1 位作者 YANG Haoran SHUAI Weiwei 《High Technology Letters》 EI CAS 2024年第1期23-30,共8页
The unsupervised multi-modal image translation is an emerging domain of computer vision whose goal is to transform an image from the source domain into many diverse styles in the target domain.However,the multi-genera... The unsupervised multi-modal image translation is an emerging domain of computer vision whose goal is to transform an image from the source domain into many diverse styles in the target domain.However,the multi-generator mechanism is employed among the advanced approaches available to model different domain mappings,which results in inefficient training of neural networks and pattern collapse,leading to inefficient generation of image diversity.To address this issue,this paper introduces a multi-modal unsupervised image translation framework that uses a generator to perform multi-modal image translation.Specifically,firstly,the domain code is introduced in this paper to explicitly control the different generation tasks.Secondly,this paper brings in the squeeze-and-excitation(SE)mechanism and feature attention(FA)module.Finally,the model integrates multiple optimization objectives to ensure efficient multi-modal translation.This paper performs qualitative and quantitative experiments on multiple non-paired benchmark image translation datasets while demonstrating the benefits of the proposed method over existing technologies.Overall,experimental results have shown that the proposed method is versatile and scalable. 展开更多
关键词 multi-modal image translation generative adversarial network(GAN) squeezeand-excitation(SE)mechanism feature attention(FA)module
下载PDF
A Lightweight Network with Dual Encoder and Cross Feature Fusion for Cement Pavement Crack Detection
2
作者 Zhong Qu Guoqing Mu Bin Yuan 《Computer Modeling in Engineering & Sciences》 SCIE EI 2024年第7期255-273,共19页
Automatic crack detection of cement pavement chiefly benefits from the rapid development of deep learning,with convolutional neural networks(CNN)playing an important role in this field.However,as the performance of cr... Automatic crack detection of cement pavement chiefly benefits from the rapid development of deep learning,with convolutional neural networks(CNN)playing an important role in this field.However,as the performance of crack detection in cement pavement improves,the depth and width of the network structure are significantly increased,which necessitates more computing power and storage space.This limitation hampers the practical implementation of crack detection models on various platforms,particularly portable devices like small mobile devices.To solve these problems,we propose a dual-encoder-based network architecture that focuses on extracting more comprehensive fracture feature information and combines cross-fusion modules and coordinated attention mechanisms formore efficient feature fusion.Firstly,we use small channel convolution to construct shallow feature extractionmodule(SFEM)to extract low-level feature information of cracks in cement pavement images,in order to obtainmore information about cracks in the shallowfeatures of images.In addition,we construct large kernel atrous convolution(LKAC)to enhance crack information,which incorporates coordination attention mechanism for non-crack information filtering,and large kernel atrous convolution with different cores,using different receptive fields to extract more detailed edge and context information.Finally,the three-stage feature map outputs from the shallow feature extraction module is cross-fused with the two-stage feature map outputs from the large kernel atrous convolution module,and the shallow feature and detailed edge feature are fully fused to obtain the final crack prediction map.We evaluate our method on three public crack datasets:DeepCrack,CFD,and Crack500.Experimental results on theDeepCrack dataset demonstrate the effectiveness of our proposed method compared to state-of-the-art crack detection methods,which achieves Precision(P)87.2%,Recall(R)87.7%,and F-score(F1)87.4%.Thanks to our lightweight crack detectionmodel,the parameter count of the model in real-world detection scenarios has been significantly reduced to less than 2M.This advancement also facilitates technical support for portable scene detection. 展开更多
关键词 Shallow feature extraction module large kernel atrous convolution dual encoder lightweight network crack detection
下载PDF
Social Robot Detection Method with Improved Graph Neural Networks
3
作者 Zhenhua Yu Liangxue Bai +1 位作者 Ou Ye Xuya Cong 《Computers, Materials & Continua》 SCIE EI 2024年第2期1773-1795,共23页
Social robot accounts controlled by artificial intelligence or humans are active in social networks,bringing negative impacts to network security and social life.Existing social robot detection methods based on graph ... Social robot accounts controlled by artificial intelligence or humans are active in social networks,bringing negative impacts to network security and social life.Existing social robot detection methods based on graph neural networks suffer from the problem of many social network nodes and complex relationships,which makes it difficult to accurately describe the difference between the topological relations of nodes,resulting in low detection accuracy of social robots.This paper proposes a social robot detection method with the use of an improved neural network.First,social relationship subgraphs are constructed by leveraging the user’s social network to disentangle intricate social relationships effectively.Then,a linear modulated graph attention residual network model is devised to extract the node and network topology features of the social relation subgraph,thereby generating comprehensive social relation subgraph features,and the feature-wise linear modulation module of the model can better learn the differences between the nodes.Next,user text content and behavioral gene sequences are extracted to construct social behavioral features combined with the social relationship subgraph features.Finally,social robots can be more accurately identified by combining user behavioral and relationship features.By carrying out experimental studies based on the publicly available datasets TwiBot-20 and Cresci-15,the suggested method’s detection accuracies can achieve 86.73%and 97.86%,respectively.Compared with the existing mainstream approaches,the accuracy of the proposed method is 2.2%and 1.35%higher on the two datasets.The results show that the method proposed in this paper can effectively detect social robots and maintain a healthy ecological environment of social networks. 展开更多
关键词 Social robot detection social relationship subgraph graph attention network feature linear modulation behavioral gene sequences
下载PDF
CFM-UNet:A Joint CNN and Transformer Network via Cross Feature Modulation for Remote Sensing Images Segmentation 被引量:3
4
作者 Min WANG Peidong WANG 《Journal of Geodesy and Geoinformation Science》 CSCD 2023年第4期40-47,共8页
The semantic segmentation methods based on CNN have made great progress,but there are still some shortcomings in the application of remote sensing images segmentation,such as the small receptive field can not effectiv... The semantic segmentation methods based on CNN have made great progress,but there are still some shortcomings in the application of remote sensing images segmentation,such as the small receptive field can not effectively capture global context.In order to solve this problem,this paper proposes a hybrid model based on ResNet50 and swin transformer to directly capture long-range dependence,which fuses features through Cross Feature Modulation Module(CFMM).Experimental results on two publicly available datasets,Vaihingen and Potsdam,are mIoU of 70.27%and 76.63%,respectively.Thus,CFM-UNet can maintain a high segmentation performance compared with other competitive networks. 展开更多
关键词 remote sensing images semantic segmentation swin transformer feature modulation module
下载PDF
RF-Net: Unsupervised Low-Light Image Enhancement Based on Retinex and Exposure Fusion
5
作者 Tian Ma Chenhui Fu +2 位作者 Jiayi Yang Jiehui Zhang Chuyang Shang 《Computers, Materials & Continua》 SCIE EI 2023年第10期1103-1122,共20页
Low-light image enhancement methods have limitations in addressing issues such as color distortion,lack of vibrancy,and uneven light distribution and often require paired training data.To address these issues,we propo... Low-light image enhancement methods have limitations in addressing issues such as color distortion,lack of vibrancy,and uneven light distribution and often require paired training data.To address these issues,we propose a two-stage unsupervised low-light image enhancement algorithm called Retinex and Exposure Fusion Network(RFNet),which can overcome the problems of over-enhancement of the high dynamic range and under-enhancement of the low dynamic range in existing enhancement algorithms.This algorithm can better manage the challenges brought about by complex environments in real-world scenarios by training with unpaired low-light images and regular-light images.In the first stage,we design a multi-scale feature extraction module based on Retinex theory,capable of extracting details and structural information at different scales to generate high-quality illumination and reflection images.In the second stage,an exposure image generator is designed through the camera response mechanism function to acquire exposure images containing more dark features,and the generated images are fused with the original input images to complete the low-light image enhancement.Experiments show the effectiveness and rationality of each module designed in this paper.And the method reconstructs the details of contrast and color distribution,outperforms the current state-of-the-art methods in both qualitative and quantitative metrics,and shows excellent performance in the real world. 展开更多
关键词 Low-light image enhancement multiscale feature extraction module exposure generator exposure fusion
下载PDF
A Single Image Derain Method Based on Residue Channel Decomposition in Edge Computing
6
作者 Yong Cheng Zexuan Yang +3 位作者 Wenjie Zhang Ling Yang Jun Wang Tingzhao Guan 《Intelligent Automation & Soft Computing》 SCIE 2023年第8期1469-1482,共14页
The numerous photos captured by low-price Internet of Things(IoT)sensors are frequently affected by meteorological factors,especially rainfall.It causes varying sizes of white streaks on the image,destroying the image... The numerous photos captured by low-price Internet of Things(IoT)sensors are frequently affected by meteorological factors,especially rainfall.It causes varying sizes of white streaks on the image,destroying the image texture and ruining the performance of the outdoor computer vision system.Existing methods utilise training with pairs of images,which is difficult to cover all scenes and leads to domain gaps.In addition,the network structures adopt deep learning to map rain images to rain-free images,failing to use prior knowledge effectively.To solve these problems,we introduce a single image derain model in edge computing that combines prior knowledge of rain patterns with the learning capability of the neural network.Specifically,the algorithm first uses Residue Channel Prior to filter out the rainfall textural features then it uses the Feature Fusion Module to fuse the original image with the background feature information.This results in a pre-processed image which is fed into Half Instance Net(HINet)to recover a high-quality rain-free image with a clear and accurate structure,and the model does not rely on any rainfall assumptions.Experimental results on synthetic and real-world datasets show that the average peak signal-to-noise ratio of the model decreases by 0.37 dB on the synthetic dataset and increases by 0.43 dB on the real-world dataset,demonstrating that a combined model reduces the gap between synthetic data and natural rain scenes,improves the generalization ability of the derain network,and alleviates the overfitting problem. 展开更多
关键词 Single image derain method edge computing residue channel prior feature fusion module
下载PDF
DB-DCAFN:dual-branch deformable cross-attention fusion network for bacterial segmentation
7
作者 Jingkun Wang Xinyu Ma +6 位作者 Long Cao Yilin Leng Zeyi Li Zihan Cheng Yuzhu Cao Xiaoping Huang Jian Zheng 《Visual Computing for Industry,Biomedicine,and Art》 EI 2023年第1期155-170,共16页
Sputum smear tests are critical for the diagnosis of respiratory diseases. Automatic segmentation of bacteria from spu-tum smear images is important for improving diagnostic efficiency. However, this remains a challen... Sputum smear tests are critical for the diagnosis of respiratory diseases. Automatic segmentation of bacteria from spu-tum smear images is important for improving diagnostic efficiency. However, this remains a challenging task owing to the high interclass similarity among different categories of bacteria and the low contrast of the bacterial edges. To explore more levels of global pattern features to promote the distinguishing ability of bacterial categories and main-tain sufficient local fine-grained features to ensure accurate localization of ambiguous bacteria simultaneously, we propose a novel dual-branch deformable cross-attention fusion network (DB-DCAFN) for accurate bacterial segmen-tation. Specifically, we first designed a dual-branch encoder consisting of multiple convolution and transformer blocks in parallel to simultaneously extract multilevel local and global features. We then designed a sparse and deformable cross-attention module to capture the semantic dependencies between local and global features, which can bridge the semantic gap and fuse features effectively. Furthermore, we designed a feature assignment fusion module to enhance meaningful features using an adaptive feature weighting strategy to obtain more accurate segmentation. We conducted extensive experiments to evaluate the effectiveness of DB-DCAFN on a clinical dataset comprising three bacterial categories: Acinetobacter baumannii, Klebsiella pneumoniae, and Pseudomonas aeruginosa. The experi-mental results demonstrate that the proposed DB-DCAFN outperforms other state-of-the-art methods and is effective at segmenting bacteria from sputum smear images. 展开更多
关键词 Bacterial segmentation Dual-branch parallel encoder Deformable cross-attention module Feature assignment fusion module
下载PDF
Automatic modulation classification using modulation fingerprint extraction 被引量:2
8
作者 NOROLAHI Jafar AZMI Paeiz AHMADI Farzaneh 《Journal of Systems Engineering and Electronics》 SCIE EI CSCD 2021年第4期799-810,共12页
An automatic method for classifying frequency shift keying(FSK),minimum shift keying(MSK),phase shift keying(PSK),quadrature amplitude modulation(QAM),and orthogonal frequency division multiplexing(OFDM)is proposed by... An automatic method for classifying frequency shift keying(FSK),minimum shift keying(MSK),phase shift keying(PSK),quadrature amplitude modulation(QAM),and orthogonal frequency division multiplexing(OFDM)is proposed by simultaneously using normality test,spectral analysis,and geometrical characteristics of in-phase-quadrature(I-Q)constellation diagram.Since the extracted features are unique for each modulation,they can be considered as a fingerprint of each modulation.We show that the proposed algorithm outperforms the previously published methods in terms of signal-to-noise ratio(SNR)and success rate.For example,the success rate of the proposed method for 64-QAM modulation at SNR=11 dB is 99%.Another advantage of the proposed method is its wide SNR range;such that the probability of classification for 16-QAM at SNR=3 dB is almost 1.The proposed method also provides a database for geometrical features of I-Q constellation diagram.By comparing and correlating the data of the provided database with the estimated I-Q diagram of the received signal,the processing gain of 4 dB is obtained.Whatever can be mentioned about the preference of the proposed algorithm are low complexity,low SNR,wide range of modulation set,and enhanced recognition at higher-order modulations. 展开更多
关键词 automatic modulation classification in-phase-quadrature(I-Q)constellation diagram spectral analysis feature based modulation classification
下载PDF
上一页 1 下一页 到第
使用帮助 返回顶部