期刊文献+
共找到308篇文章
< 1 2 16 >
每页显示 20 50 100
DCFNet:An Effective Dual-Branch Cross-Attention Fusion Network for Medical Image Segmentation
1
作者 Chengzhang Zhu Renmao Zhang +5 位作者 Yalong Xiao Beiji Zou Xian Chai Zhangzheng Yang Rong Hu Xuanchu Duan 《Computer Modeling in Engineering & Sciences》 SCIE EI 2024年第7期1103-1128,共26页
Automatic segmentation of medical images provides a reliable scientific basis for disease diagnosis and analysis.Notably,most existing methods that combine the strengths of convolutional neural networks(CNNs)and Trans... Automatic segmentation of medical images provides a reliable scientific basis for disease diagnosis and analysis.Notably,most existing methods that combine the strengths of convolutional neural networks(CNNs)and Transformers have made significant progress.However,there are some limitations in the current integration of CNN and Transformer technology in two key aspects.Firstly,most methods either overlook or fail to fully incorporate the complementary nature between local and global features.Secondly,the significance of integrating the multiscale encoder features from the dual-branch network to enhance the decoding features is often disregarded in methods that combine CNN and Transformer.To address this issue,we present a groundbreaking dual-branch cross-attention fusion network(DCFNet),which efficiently combines the power of Swin Transformer and CNN to generate complementary global and local features.We then designed the Feature Cross-Fusion(FCF)module to efficiently fuse local and global features.In the FCF,the utilization of the Channel-wise Cross-fusion Transformer(CCT)serves the purpose of aggregatingmulti-scale features,and the Feature FusionModule(FFM)is employed to effectively aggregate dual-branch prominent feature regions from the spatial perspective.Furthermore,within the decoding phase of the dual-branch network,our proposed Channel Attention Block(CAB)aims to emphasize the significance of the channel features between the up-sampled features and the features generated by the FCFmodule to enhance the details of the decoding.Experimental results demonstrate that DCFNet exhibits enhanced accuracy in segmentation performance.Compared to other state-of-the-art(SOTA)methods,our segmentation framework exhibits a superior level of competitiveness.DCFNet’s accurate segmentation of medical images can greatly assist medical professionals in making crucial diagnoses of lesion areas in advance. 展开更多
关键词 Convolutional neural networks Swin Transformer dual branch medical image segmentation feature cross fusion
下载PDF
Feature-Based Fusion of Dual Band Infrared Image Using Multiple Pulse Coupled Neural Network 被引量:1
2
作者 Yuqing He Shuaiying Wei +3 位作者 Tao Yang Weiqi Jin Mingqi Liu Xiangyang Zhai 《Journal of Beijing Institute of Technology》 EI CAS 2019年第1期129-136,共8页
To improve the quality of the infrared image and enhance the information of the object,a dual band infrared image fusion method based on feature extraction and a novel multiple pulse coupled neural network(multi-PCNN)... To improve the quality of the infrared image and enhance the information of the object,a dual band infrared image fusion method based on feature extraction and a novel multiple pulse coupled neural network(multi-PCNN)is proposed.In this multi-PCNN fusion scheme,the auxiliary PCNN which captures the characteristics of feature image extracting from the infrared image is used to modulate the main PCNN,whose input could be original infrared image.Meanwhile,to make the PCNN fusion effect consistent with the human vision system,Laplacian energy is adopted to obtain the value of adaptive linking strength in PCNN.After that,the original dual band infrared images are reconstructed by using a weight fusion rule with the fire mapping images generated by the main PCNNs to obtain the fused image.Compared to wavelet transforms,Laplacian pyramids and traditional multi-PCNNs,fusion images based on our method have more information,rich details and clear edges. 展开更多
关键词 infrared IMAGE IMAGE fusion dual Band pulse coupled NEURAL network(PCNN) feature extraction
下载PDF
Efficient Object Segmentation and Recognition Using Multi-Layer Perceptron Networks
3
作者 Aysha Naseer Nouf Abdullah Almujally +2 位作者 Saud S.Alotaibi Abdulwahab Alazeb Jeongmin Park 《Computers, Materials & Continua》 SCIE EI 2024年第1期1381-1398,共18页
Object segmentation and recognition is an imperative area of computer vision andmachine learning that identifies and separates individual objects within an image or video and determines classes or categories based on ... Object segmentation and recognition is an imperative area of computer vision andmachine learning that identifies and separates individual objects within an image or video and determines classes or categories based on their features.The proposed system presents a distinctive approach to object segmentation and recognition using Artificial Neural Networks(ANNs).The system takes RGB images as input and uses a k-means clustering-based segmentation technique to fragment the intended parts of the images into different regions and label thembased on their characteristics.Then,two distinct kinds of features are obtained from the segmented images to help identify the objects of interest.An Artificial Neural Network(ANN)is then used to recognize the objects based on their features.Experiments were carried out with three standard datasets,MSRC,MS COCO,and Caltech 101 which are extensively used in object recognition research,to measure the productivity of the suggested approach.The findings from the experiment support the suggested system’s validity,as it achieved class recognition accuracies of 89%,83%,and 90.30% on the MSRC,MS COCO,and Caltech 101 datasets,respectively. 展开更多
关键词 K-region fusion segmentation recognition feature extraction artificial neural network computer vision
下载PDF
FusionNN:A Semantic Feature Fusion Model Based on Multimodal for Web Anomaly Detection
4
作者 Li Wang Mingshan Xia +3 位作者 Hao Hu Jianfang Li Fengyao Hou Gang Chen 《Computers, Materials & Continua》 SCIE EI 2024年第5期2991-3006,共16页
With the rapid development of the mobile communication and the Internet,the previous web anomaly detectionand identificationmodels were built relying on security experts’empirical knowledge and attack features.Althou... With the rapid development of the mobile communication and the Internet,the previous web anomaly detectionand identificationmodels were built relying on security experts’empirical knowledge and attack features.Althoughthis approach can achieve higher detection performance,it requires huge human labor and resources to maintainthe feature library.In contrast,semantic feature engineering can dynamically discover new semantic featuresand optimize feature selection by automatically analyzing the semantic information contained in the data itself,thus reducing dependence on prior knowledge.However,current semantic features still have the problem ofsemantic expression singularity,as they are extracted from a single semantic mode such as word segmentation,character segmentation,or arbitrary semantic feature extraction.This paper extracts features of web requestsfrom dual semantic granularity,and proposes a semantic feature fusion method to solve the above problems.Themethod first preprocesses web requests,and extracts word-level and character-level semantic features of URLs viaconvolutional neural network(CNN),respectively.By constructing three loss functions to reduce losses betweenfeatures,labels and categories.Experiments on the HTTP CSIC 2010,Malicious URLs and HttpParams datasetsverify the proposedmethod.Results show that compared withmachine learning,deep learningmethods and BERTmodel,the proposed method has better detection performance.And it achieved the best detection rate of 99.16%in the dataset HttpParams. 展开更多
关键词 feature fusion web anomaly detection MULTIMODAL convolutional neural network(CNN) semantic feature extraction
下载PDF
Feature Fusion Multi_XMNet Convolution Neural Network for Clothing Image Classification 被引量:2
5
作者 ZHOU Honglei PENG Zhifei +1 位作者 TAO Ran ZHANG Lu 《Journal of Donghua University(English Edition)》 CAS 2021年第6期519-526,共8页
Faced with the massive amount of online shopping clothing images,how to classify them quickly and accurately is a challenging task in image classification.In this paper,we propose a novel method,named Multi_XMNet,to s... Faced with the massive amount of online shopping clothing images,how to classify them quickly and accurately is a challenging task in image classification.In this paper,we propose a novel method,named Multi_XMNet,to solve the clothing images classification problem.The proposed method mainly consists of two convolution neural network(CNN)branches.One branch extracts multiscale features from the whole expressional image by Multi_X which is designed by improving the Xception network,while the other extracts attention mechanism features from the whole expressional image by MobileNetV3-small network.Both multiscale and attention mechanism features are aggregated before making classification.Additionally,in the training stage,global average pooling(GAP),convolutional layers,and softmax classifiers are used instead of the fully connected layer to classify the final features,which speed up model training and alleviate the problem of overfitting caused by too many parameters.Experimental comparisons are made in the public DeepFashion dataset.The experimental results show that the classification accuracy of this method is 95.38%,which is better than InceptionV3,Xception and InceptionV3_Xception by 5.58%,3.32%,and 2.22%,respectively.The proposed Multi_XMNet image classification model can help enterprises and researchers in the field of clothing e-commerce to automaticly,efficiently and accurately classify massive clothing images. 展开更多
关键词 feature extraction feature fusion multiscale feature convolution neural network(CNN) clothing image classification
下载PDF
Residual Feature Attentional Fusion Network for Lightweight Chest CT Image Super-Resolution 被引量:1
6
作者 Kun Yang Lei Zhao +4 位作者 Xianghui Wang Mingyang Zhang Linyan Xue Shuang Liu Kun Liu 《Computers, Materials & Continua》 SCIE EI 2023年第6期5159-5176,共18页
The diagnosis of COVID-19 requires chest computed tomography(CT).High-resolution CT images can provide more diagnostic information to help doctors better diagnose the disease,so it is of clinical importance to study s... The diagnosis of COVID-19 requires chest computed tomography(CT).High-resolution CT images can provide more diagnostic information to help doctors better diagnose the disease,so it is of clinical importance to study super-resolution(SR)algorithms applied to CT images to improve the reso-lution of CT images.However,most of the existing SR algorithms are studied based on natural images,which are not suitable for medical images;and most of these algorithms improve the reconstruction quality by increasing the network depth,which is not suitable for machines with limited resources.To alleviate these issues,we propose a residual feature attentional fusion network for lightweight chest CT image super-resolution(RFAFN).Specifically,we design a contextual feature extraction block(CFEB)that can extract CT image features more efficiently and accurately than ordinary residual blocks.In addition,we propose a feature-weighted cascading strategy(FWCS)based on attentional feature fusion blocks(AFFB)to utilize the high-frequency detail information extracted by CFEB as much as possible via selectively fusing adjacent level feature information.Finally,we suggest a global hierarchical feature fusion strategy(GHFFS),which can utilize the hierarchical features more effectively than dense concatenation by progressively aggregating the feature information at various levels.Numerous experiments show that our method performs better than most of the state-of-the-art(SOTA)methods on the COVID-19 chest CT dataset.In detail,the peak signal-to-noise ratio(PSNR)is 0.11 dB and 0.47 dB higher on CTtest1 and CTtest2 at×3 SR compared to the suboptimal method,but the number of parameters and multi-adds are reduced by 22K and 0.43G,respectively.Our method can better recover chest CT image quality with fewer computational resources and effectively assist in COVID-19. 展开更多
关键词 SUPER-RESOLUTION COVID-19 chest CT lightweight network contextual feature extraction attentional feature fusion
下载PDF
Cycle GAN-MF:A Cycle-consistent Generative Adversarial Network Based on Multifeature Fusion for Pedestrian Re-recognition 被引量:3
7
作者 Yongqi Fan Li Hang Botong Sun 《IJLAI Transactions on Science and Engineering》 2024年第1期38-45,共8页
In pedestrian re-recognition,the traditional pedestrian re-recognition method will be affected by the changes of background,veil,clothing and so on,which will make the recognition effect decline.In order to reduce the... In pedestrian re-recognition,the traditional pedestrian re-recognition method will be affected by the changes of background,veil,clothing and so on,which will make the recognition effect decline.In order to reduce the impact of background,veil,clothing and other changes on the recognition effect,this paper proposes a pedestrian re-recognition method based on the cycle-consistent generative adversarial network and multifeature fusion.By comparing the measured distance between two pedestrians,pedestrian re-recognition is accomplished.Firstly,this paper uses Cycle GAN to transform and expand the data set,so as to reduce the influence of pedestrian posture changes as much as possible.The method consists of two branches:global feature extraction and local feature extraction.Then the global feature and local feature are fused.The fused features are used for comparison measurement learning,and the similarity scores are calculated to sort the samples.A large number of experimental results on large data sets CUHK03 and VIPER show that this new method reduces the influence of background,veil,clothing and other changes on the recognition effect. 展开更多
关键词 Pedestrian re-recognition Cycle-consistent generative adversarial network Multifeature fusion Global feature extraction Local feature extraction
原文传递
Recognition of weeds at asparagus fields using multi-feature fusion and backpropagation neural network 被引量:2
8
作者 Yafei Wang Xiaodong Zhang +3 位作者 Guoxin Ma Xiaoxue Du Naila Shaheen Hanping Mao 《International Journal of Agricultural and Biological Engineering》 SCIE EI CAS 2021年第4期190-198,共9页
In order to solve the problem of low recognition rates of weeds by a single feature,a method was proposed in this study to identify weeds in Asparagus(Asparagus officinalis L.)field using multi-feature fusion and back... In order to solve the problem of low recognition rates of weeds by a single feature,a method was proposed in this study to identify weeds in Asparagus(Asparagus officinalis L.)field using multi-feature fusion and backpropagation neural network(BPNN).A total of 382 images of weeds competing with asparagus growth were collected,including 135 of Cirsium arvense(L.)Scop.,138 of Conyza sumatrensis(Retz.)E.Walker,and 109 of Calystegia hederacea Wall.The grayscale images were extracted from the RGB images of weeds using the 2G-R-B factor.Threshold segmentation of the grayscale image of weeds was applied using Otsu method.Then the internal holes of the leaves were filled through the expansion and corrosion morphological operations,and other interference targets were removed to obtain the binary image.The foreground image was obtained by masking the binary image and the RGB image.Then,the color moment algorithm was used to extract weeds color feature,the gray level co-occurrence matrix and the Local Binary Pattern(LBP)algorithm was used to extract weeds texture features,and seven Hu invariant moment features and the roundness and slenderness ratio of weeds were extracted as their shape features.According to the shape,color,texture,and fusion features of the test samples,a weed identification model was built.The test results showed that the recognition rate of Cirsium arvense(L.)Scop.,Calystegia hederacea Wall.and Conyza sumatrensis(Retz.)E.Walker were 82.72%(color feature),72.41%(shape feature),86.73%(texture feature)and 93.51%(fusion feature),respectively.Therefore,this method can provide a reference for the study of weeds identification in the asparagus field. 展开更多
关键词 weeds recognition image processing feature extraction multi-feature fusion BP neural network asparagus field
原文传递
基于注意力机制与光照感知网络的红外与可见光图像融合 被引量:1
9
作者 杨艳春 闫岩 王可 《西南交通大学学报》 EI CSCD 北大核心 2024年第5期1204-1214,共11页
部分图像融合方法未充分考虑图像环境的光照状况,导致融合图像中出现红外目标亮度不足以及整体画面亮度较低,从而影响纹理细节的清晰度.为解决上述问题,提出一种基于注意力机制与光照感知网络相结合的红外与可见光图像融合算法.首先,在... 部分图像融合方法未充分考虑图像环境的光照状况,导致融合图像中出现红外目标亮度不足以及整体画面亮度较低,从而影响纹理细节的清晰度.为解决上述问题,提出一种基于注意力机制与光照感知网络相结合的红外与可见光图像融合算法.首先,在训练融合网络之前利用光照感知网络计算当前场景是日间或夜间的概率,将其运用至融合网络损失函数中,用以指导融合网络训练;然后,在网络的特征提取部分采用空间注意力机制和深度可分离卷积对源图像进行特征提取,得到空间显著信息后,输入卷积神经网络(CNN)以提取深度特征;最后,将深度特征信息进行拼接用于图像重建,进而得到最终的融合图像.实验结果表明:本文方法的互信息(MI)、视觉保真度(VIF)、平均梯度(AG)、融合质量(Qabf)与空间频率(SF)较对比方法分别平均提高39.33%、11.29%、26.27%、47.11%和39.01%;融合后的图像能够有效保留红外目标亮度,且包含丰富的纹理细节信息. 展开更多
关键词 图像融合 注意力机制 卷积神经网络 红外特征提取 深度学习
下载PDF
多模态融合的三维目标检测方法研究
10
作者 田枫 宗内丽 +5 位作者 刘芳 卢圆圆 刘超 姜文文 赵玲 韩玉祥 《计算机工程与应用》 CSCD 北大核心 2024年第13期113-123,共11页
针对点云稀疏性与无序性导致基于纯点云的检测算法容易出现远小目标漏检和误检的问题,提出一种融合图像特征与点云体素特征的多模态三维目标检测算法。在图像特征提取阶段,提出一种轻量级深度残差网络,减少图像特征通道数,使其与点云体... 针对点云稀疏性与无序性导致基于纯点云的检测算法容易出现远小目标漏检和误检的问题,提出一种融合图像特征与点云体素特征的多模态三维目标检测算法。在图像特征提取阶段,提出一种轻量级深度残差网络,减少图像特征通道数,使其与点云体素特征相一致,提高点云和图像特征的融合能力;在体素特征与图像特征融合阶段,提出一种双次特征融合网络,在保留原始体素特征结构信息的基础上将图像特征和体素特征进行融合,使点云具备丰富的语义信息,提高远小目标检测精度。在KITTI数据集上实验结果显示,与基线模型相比,对小汽车、骑行者与行人的3D平均检测精度分别提高了0.76个百分点、2.30个百分点、3.43个百分点。实验结果验证了所提方法对于解决远小目标误检和漏检问题的有效性。 展开更多
关键词 三维目标检测 深度残差网络 体素特征 图像特征 特征融合 双次特征融合网络
下载PDF
基于跨模态注意力融合的煤炭异物检测方法 被引量:1
11
作者 曹现刚 李虎 +3 位作者 王鹏 吴旭东 向敬芳 丁文韬 《工矿自动化》 CSCD 北大核心 2024年第1期57-65,共9页
为解决原煤智能化洗选过程中煤流中夹杂的异物对比度低、相互遮挡导致异物图像检测时特征提取不充分的问题,提出了一种基于跨模态注意力融合的煤炭异物检测方法。通过引入Depth图像构建RGB图像与Depth图像的双特征金字塔网络(DFPN),采... 为解决原煤智能化洗选过程中煤流中夹杂的异物对比度低、相互遮挡导致异物图像检测时特征提取不充分的问题,提出了一种基于跨模态注意力融合的煤炭异物检测方法。通过引入Depth图像构建RGB图像与Depth图像的双特征金字塔网络(DFPN),采用浅层的特征提取策略提取Depth图像的低级特征,用深度边缘与深度纹理等基础特征辅助RGB图像深层特征,以有效获得2种特征的互补信息,从而丰富异物特征的空间与边缘信息,提高检测精度;构建了基于坐标注意力与改进空间注意力的跨模态注意力融合模块(CAFM),以协同优化并融合RGB特征与Depth特征,增强网络对特征图中被遮挡异物可见部分的关注度,提高被遮挡异物检测精度;使用区域卷积神经网络(R-CNN)输出煤炭异物的分类、回归与分割结果。实验结果表明:在检测精度方面,该方法的AP相较两阶段模型中较优的Mask transfiner高3.9%;在检测效率方面,该方法的单帧检测时间为110.5 ms,能够满足异物检测实时性需求。基于跨模态注意力融合的煤炭异物检测方法能够以空间特征辅助色彩、形状与纹理等特征,准确识别煤炭异物之间及煤炭异物与输送带之间的差异,从而有效提高对复杂特征异物的检测精度,减少误检、漏检现象,实现复杂特征下煤炭异物的精确检测与像素级分割。 展开更多
关键词 煤炭异物检测 实例分割 双特征金字塔网络 跨模态注意力融合 Depth图像 坐标注意力 改进空间注意力
下载PDF
基于三分支对抗学习和补偿注意力的红外和可见光图像融合
12
作者 邸敬 任莉 +2 位作者 刘冀钊 郭文庆 廉敬 《红外技术》 CSCD 北大核心 2024年第5期510-521,共12页
针对现有深度学习图像融合方法依赖卷积提取特征,并未考虑源图像全局特征,融合结果容易产生纹理模糊、对比度低等问题,本文提出一种基于三分支对抗学习和补偿注意力的红外和可见光图像融合方法。首先,生成器网络采用密集块和补偿注意力... 针对现有深度学习图像融合方法依赖卷积提取特征,并未考虑源图像全局特征,融合结果容易产生纹理模糊、对比度低等问题,本文提出一种基于三分支对抗学习和补偿注意力的红外和可见光图像融合方法。首先,生成器网络采用密集块和补偿注意力机制构建局部-全局三分支提取特征信息。然后,利用通道特征和空间特征变化构建补偿注意力机制提取全局信息,更进一步提取红外目标和可见光细节表征。其次,设计聚焦双对抗鉴别器,以确定融合结果和源图像之间的相似分布。最后,选用公开数据集TNO和RoadScene进行实验并与其他9种具有代表性的图像融合方法进行对比,本文提出的方法不仅获得纹理细节更清晰、对比度更好的融合结果,而且客观度量指标优于其他先进方法。 展开更多
关键词 红外可见光图像融合 局部-全局三分支 局部特征提取 补偿注意力机制 对抗学习 聚焦双对抗鉴别器
下载PDF
卷积神经网提取特征的红外与可见光图像融合研究
13
作者 郑晓东 郑业爽 栾国森 《激光杂志》 CAS 北大核心 2024年第5期133-138,共6页
当前红外与可见光图像融合存在一些难题,导致前红外与可见光图像精度低,误差大,而且前红外与可见光图像融合效率低,为了解决当前红外与可见光图像过程存在的问题,设计了基于卷积神经网提取特征的红外与可见光图像融合方法。首先分别采... 当前红外与可见光图像融合存在一些难题,导致前红外与可见光图像精度低,误差大,而且前红外与可见光图像融合效率低,为了解决当前红外与可见光图像过程存在的问题,设计了基于卷积神经网提取特征的红外与可见光图像融合方法。首先分别采集对象的红外图像和可见光图像,并对原始图像去噪等预处理,改善图像的质量,然后采用卷积神经网络提取红外与可见光图像融合特征,根据特征得到红外与可见光图像融合结果,最后进行了仿真实验,结果表明本方法红外与可见光图像的融合结果的融合比率提高了0.24,平均梯度值提升了0.22,图像融合质量更高。 展开更多
关键词 卷积神经网络 提取特征 融合比率 仿真测试
下载PDF
联合多连接特征编解码与小波池化的轻量级语义分割
14
作者 易清明 王渝 +1 位作者 石敏 骆爱文 《电子科技大学学报》 EI CAS CSCD 北大核心 2024年第3期366-375,共10页
语义分割是当前场景理解领域的基础技术之一。现存的语义分割网络通常结构复杂、参数量大、图像特征信息损失过多和计算效率低。针对以上问题,基于编-解码器框架和离散小波变换,设计了一个联合多连接特征编解码与小波池化的轻量级语义... 语义分割是当前场景理解领域的基础技术之一。现存的语义分割网络通常结构复杂、参数量大、图像特征信息损失过多和计算效率低。针对以上问题,基于编-解码器框架和离散小波变换,设计了一个联合多连接特征编解码与小波池化的轻量级语义分割网络MLWP-Net(Multi-Link Wavelet-Pooled Network),在编码阶段利用多连接策略并结合深度可分离卷积、空洞卷积和通道压缩设计了轻量级特征提取瓶颈结构,并设计了低频混合小波池化操作替代传统的下采样操作,有效降低编码过程造成的信息丢失;在解码阶段,设计了多分支并行空洞卷积解码器以融合多级特征并行实现图像分辨率的恢复。实验结果表明,MLWP-Net仅以0.74 MB的参数量在数据集Cityscapes和CamVid上分别达到74.1%和68.2%mIoU的分割精度,验证了该算法的有效性。 展开更多
关键词 实时语义分割 轻量级神经网络 多连接特征融合 小波池化 多分支空洞卷积
下载PDF
基于多源信息融合的RBF神经网络室内可见光定位算法
15
作者 王琪 孟祥艳 赵黎 《光通信技术》 北大核心 2024年第2期30-35,共6页
针对基于接收信号强度(RSS)的定位技术易受环境干扰而导致定位精度不高和稳定性较差的问题,提出了一种基于多源信息融合的径向基函数(RBF)神经网络室内可见光定位算法。通过将图像的颜色矩特征与RSS矩特征融合,构建指纹库,并采用RBF神... 针对基于接收信号强度(RSS)的定位技术易受环境干扰而导致定位精度不高和稳定性较差的问题,提出了一种基于多源信息融合的径向基函数(RBF)神经网络室内可见光定位算法。通过将图像的颜色矩特征与RSS矩特征融合,构建指纹库,并采用RBF神经网络进行预测,实现了图像与RSS之间的优势互补,最后对定位算法进行了验证。实验结果表明,经过优化的多源信息融合定位算法较单一RSS定位算法的定位精度提高了9.4%。 展开更多
关键词 可见光 室内定位 多源信息融合 颜色矩 神经网络 径向基函数 特征提取
下载PDF
MCFNet:融合上下文信息的多尺度视网膜动静脉分类网络
16
作者 崔颖 朱佳 +2 位作者 高山 陈立伟 张广 《应用科技》 CAS 2024年第2期105-111,共7页
针对由于血管类间具有强相似性造成的动静脉错误分类问题,提出了一种新的融合上下文信息的多尺度视网膜动静脉分类网络(multi-scale retinal artery and vein classification network,MCFNet),该网络使用多尺度特征(multi-scale feature... 针对由于血管类间具有强相似性造成的动静脉错误分类问题,提出了一种新的融合上下文信息的多尺度视网膜动静脉分类网络(multi-scale retinal artery and vein classification network,MCFNet),该网络使用多尺度特征(multi-scale feature,MSF)提取模块及高效的全局上下文信息融合(efficient global contextual information aggregation,EGCA)模块结合U型分割网络进行动静脉分类,抑制了倾向于背景的特征并增强了血管的边缘、交点和末端特征,解决了段内动静脉错误分类问题。此外,在U型网络的解码器部分加入3层深度监督,使浅层信息得到充分训练,避免梯度消失,优化训练过程。在2个公开的眼底图像数据集(DRIVE-AV,LES-AV)上,与3种现有网络进行方法对比,该模型的F1评分分别提高了2.86、1.92、0.81个百分点,灵敏度分别提高了4.27、2.43、1.21个百分点,结果表明所提出的模型能够很好地解决动静脉分类错误的问题。 展开更多
关键词 多类分割 动静脉分类 视网膜图像 多尺度特征提取 血管分割 全局信息融合 卷积神经网络 深度监督
下载PDF
结合Swin Transformer的多尺度遥感图像变化检测研究
17
作者 刘丽 张起凡 +1 位作者 白宇昂 黄凯烨 《图学学报》 CSCD 北大核心 2024年第5期941-956,共16页
由于地物信息的复杂性及变化检测数据的多元性,遥感图像特征提取的充分性和有效性难以得到保证,导致变化检测方法获取的检测结果可靠性较低。虽然卷积神经网络(CNN)凭借有效提取语义特征的优势,被广泛应用于遥感领域的变化检测之中,但... 由于地物信息的复杂性及变化检测数据的多元性,遥感图像特征提取的充分性和有效性难以得到保证,导致变化检测方法获取的检测结果可靠性较低。虽然卷积神经网络(CNN)凭借有效提取语义特征的优势,被广泛应用于遥感领域的变化检测之中,但卷积操作固有的局部性导致感受野受限,无法捕获时空上的全局信息以至于特征空间对中远距离依赖关系的建模受限。为捕获远距离的语义依赖,提取深层全局语义特征,设计了一种基于Swin Transformer的多尺度特征融合网络SwinChangeNet。首先,SwinChangeNet采用孪生的多级Swin Transformer特征编码器进行远距离上下文建模;其次,编码器中引入特征差异提取模块,计算不同尺度下变化前后的多级特征差异,再通过自适应融合层将多尺度特征图进行融合;最后,引入残差连接和通道注意力机制对融合后的特征信息进行解码,从而生成完整准确的变化图。在CDD和CD_Data_GZ 2个公开数据集上分别与7种经典和前沿变化检测方法进行比较,CDD数据集中本文模型的性能最优,相比于性能第二的模型,F1分数提高了1.11%,精确率提高了2.38%。CD_Data_GZ数据集中本文模型的性能最优,相比于性能第二的模型,F1分数、精确率和召回率分别提高了4.78%,4.32%,4.09%,提升幅度较大。对比实验结果证明了该模型具有更好的检测效果。在消融实验中也证实了模型中各个改进模块的稳定性和有效性。本文模型针对遥感图像变化检测任务,引入了Swin Transformer结构,使网络可以对遥感图像的局部特征和全局特征进行更有效地编码,让检测结果更加准确,同时保证网络在地物要素种类繁多的数据集上容易收敛。 展开更多
关键词 变化检测 孪生网络 Swin Transformer 多尺度特征融合 注意力机制 特征差异提取
下载PDF
多尺度融合与双输出U-Net网络的行人重识别
18
作者 胡庆 《计算机工程》 CAS CSCD 北大核心 2024年第6期102-109,共8页
受行人姿态的多变性和行人被遮挡等因素的影响,行人重识别模型难以提取行人关键特征。为增强模型的特征表达能力,提出一种基于多尺度融合与双输出U-Net网络的行人重识别方法,旨在解决现有方法中难以提取行人关键特征、特征表达能力较低... 受行人姿态的多变性和行人被遮挡等因素的影响,行人重识别模型难以提取行人关键特征。为增强模型的特征表达能力,提出一种基于多尺度融合与双输出U-Net网络的行人重识别方法,旨在解决现有方法中难以提取行人关键特征、特征表达能力较低的问题。首先,提出多尺度融合的双输出U-Net网络,并对输出特征进行欧氏距离和散度距离约束;其次,设计联合损失函数,解决生成对抗网络在训练过程中不易收敛的问题,提高训练过程的收敛速度。在3个公共基准数据集上的仿真实验结果表明,相比经典特征提取网络,所提特征提取网络的平均精度均值(mAP)提升超过10%,所提行人重识别方法相比主流方法的mAP提高约2%,该方法能够增强模型的特征表达能力,提高行人重识别的准确率。 展开更多
关键词 行人重识别 生成对抗网络 特征提取 多尺度融合 联合约束
下载PDF
多层特征融合与语义增强的盲图像质量评价
19
作者 赵文清 许丽娇 +1 位作者 陈昊阳 李梦伟 《智能系统学报》 CSCD 北大核心 2024年第1期132-141,共10页
针对现有盲图像质量评价算法在面对真实失真图像时性能较差的问题,本文提出多层特征融合和语义信息增强相结合的无参考图像质量评价算法。提取图像的局部和全局失真特征,利用特征融合模块对特征进行多层融合;利用多层扩张卷积增强语义信... 针对现有盲图像质量评价算法在面对真实失真图像时性能较差的问题,本文提出多层特征融合和语义信息增强相结合的无参考图像质量评价算法。提取图像的局部和全局失真特征,利用特征融合模块对特征进行多层融合;利用多层扩张卷积增强语义信息,进而指导失真图像到质量分数的映射过程;考虑预测分数和主观分数之间的相对排名关系,对L_(1)损失函数和三元组排名损失函数进行融合,构建新的损失函数L_(mix)。为了验证本文方法的有效性,在野生图像质量挑战数据集上进行了验证和对比实验,该算法的斯皮尔曼等级相关系数与皮尔逊线性相关系数指标相比原算法分别提升2.3%和2.3%;在康斯坦茨真实图像质量数据数据集和野生图像质量挑战数据集上进行了跨数据集实验,该算法在面对真实失真图像时表现出了良好的泛化性能。 展开更多
关键词 深度学习 图像质量 卷积神经网络 特征提取 通道注意力结构 多层次特征融合 扩张卷积 三元组损失函数
下载PDF
基于编解码网络的SAR影像建筑物提取
20
作者 苗国英 王慧琴 张恩伟 《遥感信息》 CSCD 北大核心 2024年第2期127-134,共8页
针对合成孔径雷达(synthetic aperture radar,SAR)遥感图像中出现的散斑噪声和几何畸变对建筑物提取过程造成干扰的问题,提出了一种多尺度特征注意力融合(multi-scale feature attention fusion,MSFAF)网络。首先,结合深度神经网络和SA... 针对合成孔径雷达(synthetic aperture radar,SAR)遥感图像中出现的散斑噪声和几何畸变对建筑物提取过程造成干扰的问题,提出了一种多尺度特征注意力融合(multi-scale feature attention fusion,MSFAF)网络。首先,结合深度神经网络和SAR图像的优势,在深层处设计了一个空间注意力融合(spatial attention fusion,SAF)模块,来整合不同层次的特征以及关注重要的空间信息。然后,利用不同尺度的卷积核以及对通道信息的转换,提出了一个多尺度细节提取(multi-scale detail extraction,MSDE)模块用于提取不同尺度的特征信息和重新分配通道信息,有利于缓解散斑噪声的干扰问题。实验证明了所提方法在SAR图像建筑物提取中取得了比其他现存方法更加优秀的性能。 展开更多
关键词 SAR遥感图像 建筑物提取 多尺度特征 注意力融合 神经网络
下载PDF
上一页 1 2 16 下一页 到第
使用帮助 返回顶部