期刊文献+
共找到486篇文章
< 1 2 25 >
每页显示 20 50 100
A Hand Features Based Fusion Recognition Network with Enhancing Multi-Modal Correlation
1
作者 Wei Wu Yuan Zhang +2 位作者 Yunpeng Li Chuanyang Li YanHao 《Computer Modeling in Engineering & Sciences》 SCIE EI 2024年第7期537-555,共19页
Fusing hand-based features in multi-modal biometric recognition enhances anti-spoofing capabilities.Additionally,it leverages inter-modal correlation to enhance recognition performance.Concurrently,the robustness and ... Fusing hand-based features in multi-modal biometric recognition enhances anti-spoofing capabilities.Additionally,it leverages inter-modal correlation to enhance recognition performance.Concurrently,the robustness and recognition performance of the system can be enhanced through judiciously leveraging the correlation among multimodal features.Nevertheless,two issues persist in multi-modal feature fusion recognition:Firstly,the enhancement of recognition performance in fusion recognition has not comprehensively considered the inter-modality correlations among distinct modalities.Secondly,during modal fusion,improper weight selection diminishes the salience of crucial modal features,thereby diminishing the overall recognition performance.To address these two issues,we introduce an enhanced DenseNet multimodal recognition network founded on feature-level fusion.The information from the three modalities is fused akin to RGB,and the input network augments the correlation between modes through channel correlation.Within the enhanced DenseNet network,the Efficient Channel Attention Network(ECA-Net)dynamically adjusts the weight of each channel to amplify the salience of crucial information in each modal feature.Depthwise separable convolution markedly reduces the training parameters and further enhances the feature correlation.Experimental evaluations were conducted on four multimodal databases,comprising six unimodal databases,including multispectral palmprint and palm vein databases from the Chinese Academy of Sciences.The Equal Error Rates(EER)values were 0.0149%,0.0150%,0.0099%,and 0.0050%,correspondingly.In comparison to other network methods for palmprint,palm vein,and finger vein fusion recognition,this approach substantially enhances recognition performance,rendering it suitable for high-security environments with practical applicability.The experiments in this article utilized amodest sample database comprising 200 individuals.The subsequent phase involves preparing for the extension of the method to larger databases. 展开更多
关键词 BIOMETRICS multi-modal CORRELATION deep learning feature-level fusion
下载PDF
A Comprehensive Survey on Deep Learning Multi-Modal Fusion:Methods,Technologies and Applications
2
作者 Tianzhe Jiao Chaopeng Guo +2 位作者 Xiaoyue Feng Yuming Chen Jie Song 《Computers, Materials & Continua》 SCIE EI 2024年第7期1-35,共35页
Multi-modal fusion technology gradually become a fundamental task in many fields,such as autonomous driving,smart healthcare,sentiment analysis,and human-computer interaction.It is rapidly becoming the dominant resear... Multi-modal fusion technology gradually become a fundamental task in many fields,such as autonomous driving,smart healthcare,sentiment analysis,and human-computer interaction.It is rapidly becoming the dominant research due to its powerful perception and judgment capabilities.Under complex scenes,multi-modal fusion technology utilizes the complementary characteristics of multiple data streams to fuse different data types and achieve more accurate predictions.However,achieving outstanding performance is challenging because of equipment performance limitations,missing information,and data noise.This paper comprehensively reviews existing methods based onmulti-modal fusion techniques and completes a detailed and in-depth analysis.According to the data fusion stage,multi-modal fusion has four primary methods:early fusion,deep fusion,late fusion,and hybrid fusion.The paper surveys the three majormulti-modal fusion technologies that can significantly enhance the effect of data fusion and further explore the applications of multi-modal fusion technology in various fields.Finally,it discusses the challenges and explores potential research opportunities.Multi-modal tasks still need intensive study because of data heterogeneity and quality.Preserving complementary information and eliminating redundant information between modalities is critical in multi-modal technology.Invalid data fusion methods may introduce extra noise and lead to worse results.This paper provides a comprehensive and detailed summary in response to these challenges. 展开更多
关键词 multi-modal fusion REPRESENTATION TRANSLATION ALIGNMENT deep learning comparative analysis
下载PDF
PowerDetector:Malicious PowerShell Script Family Classification Based on Multi-Modal Semantic Fusion and Deep Learning 被引量:1
3
作者 Xiuzhang Yang Guojun Peng +2 位作者 Dongni Zhang Yuhang Gao Chenguang Li 《China Communications》 SCIE CSCD 2023年第11期202-224,共23页
Power Shell has been widely deployed in fileless malware and advanced persistent threat(APT)attacks due to its high stealthiness and live-off-theland technique.However,existing works mainly focus on deobfuscation and ... Power Shell has been widely deployed in fileless malware and advanced persistent threat(APT)attacks due to its high stealthiness and live-off-theland technique.However,existing works mainly focus on deobfuscation and malicious detection,lacking the malicious Power Shell families classification and behavior analysis.Moreover,the state-of-the-art methods fail to capture fine-grained features and semantic relationships,resulting in low robustness and accuracy.To this end,we propose Power Detector,a novel malicious Power Shell script detector based on multimodal semantic fusion and deep learning.Specifically,we design four feature extraction methods to extract key features from character,token,abstract syntax tree(AST),and semantic knowledge graph.Then,we intelligently design four embeddings(i.e.,Char2Vec,Token2Vec,AST2Vec,and Rela2Vec) and construct a multi-modal fusion algorithm to concatenate feature vectors from different views.Finally,we propose a combined model based on transformer and CNN-Bi LSTM to implement Power Shell family detection.Our experiments with five types of Power Shell attacks show that PowerDetector can accurately detect various obfuscated and stealth PowerShell scripts,with a 0.9402 precision,a 0.9358 recall,and a 0.9374 F1-score.Furthermore,through singlemodal and multi-modal comparison experiments,we demonstrate that PowerDetector’s multi-modal embedding and deep learning model can achieve better accuracy and even identify more unknown attacks. 展开更多
关键词 deep learning malicious family detection multi-modal semantic fusion POWERSHELL
下载PDF
Multi-Modal Military Event Extraction Based on Knowledge Fusion
4
作者 Yuyuan Xiang Yangli Jia +1 位作者 Xiangliang Zhang Zhenling Zhang 《Computers, Materials & Continua》 SCIE EI 2023年第10期97-114,共18页
Event extraction stands as a significant endeavor within the realm of information extraction,aspiring to automatically extract structured event information from vast volumes of unstructured text.Extracting event eleme... Event extraction stands as a significant endeavor within the realm of information extraction,aspiring to automatically extract structured event information from vast volumes of unstructured text.Extracting event elements from multi-modal data remains a challenging task due to the presence of a large number of images and overlapping event elements in the data.Although researchers have proposed various methods to accomplish this task,most existing event extraction models cannot address these challenges because they are only applicable to text scenarios.To solve the above issues,this paper proposes a multi-modal event extraction method based on knowledge fusion.Specifically,for event-type recognition,we use a meticulous pipeline approach that integrates multiple pre-trained models.This approach enables a more comprehensive capture of the multidimensional event semantic features present in military texts,thereby enhancing the interconnectedness of information between trigger words and events.For event element extraction,we propose a method for constructing a priori templates that combine event types with corresponding trigger words.This approach facilitates the acquisition of fine-grained input samples containing event trigger words,thus enabling the model to understand the semantic relationships between elements in greater depth.Furthermore,a fusion method for spatial mapping of textual event elements and image elements is proposed to reduce the category number overload and effectively achieve multi-modal knowledge fusion.The experimental results based on the CCKS 2022 dataset show that our method has achieved competitive results,with a comprehensive evaluation value F1-score of 53.4%for the model.These results validate the effectiveness of our method in extracting event elements from multi-modal data. 展开更多
关键词 Event extraction multi-modal knowledge fusion pre-trained models
下载PDF
Robust Symmetry Prediction with Multi-Modal Feature Fusion for Partial Shapes
5
作者 Junhua Xi Kouquan Zheng +3 位作者 Yifan Zhong Longjiang Li Zhiping Cai Jinjing Chen 《Intelligent Automation & Soft Computing》 SCIE 2023年第3期3099-3111,共13页
In geometry processing,symmetry research benefits from global geo-metric features of complete shapes,but the shape of an object captured in real-world applications is often incomplete due to the limited sensor resoluti... In geometry processing,symmetry research benefits from global geo-metric features of complete shapes,but the shape of an object captured in real-world applications is often incomplete due to the limited sensor resolution,single viewpoint,and occlusion.Different from the existing works predicting symmetry from the complete shape,we propose a learning approach for symmetry predic-tion based on a single RGB-D image.Instead of directly predicting the symmetry from incomplete shapes,our method consists of two modules,i.e.,the multi-mod-al feature fusion module and the detection-by-reconstruction module.Firstly,we build a channel-transformer network(CTN)to extract cross-fusion features from the RGB-D as the multi-modal feature fusion module,which helps us aggregate features from the color and the depth separately.Then,our self-reconstruction net-work based on a 3D variational auto-encoder(3D-VAE)takes the global geo-metric features as input,followed by a prediction symmetry network to detect the symmetry.Our experiments are conducted on three public datasets:ShapeNet,YCB,and ScanNet,we demonstrate that our method can produce reliable and accurate results. 展开更多
关键词 Symmetry prediction multi-modal feature fusion partial shapes
下载PDF
Multimodal Sentiment Analysis Using BiGRU and Attention-Based Hybrid Fusion Strategy 被引量:1
6
作者 Zhizhong Liu Bin Zhou +1 位作者 Lingqiang Meng Guangyu Huang 《Intelligent Automation & Soft Computing》 SCIE 2023年第8期1963-1981,共19页
Recently,multimodal sentiment analysis has increasingly attracted attention with the popularity of complementary data streams,which has great potential to surpass unimodal sentiment analysis.One challenge of multimoda... Recently,multimodal sentiment analysis has increasingly attracted attention with the popularity of complementary data streams,which has great potential to surpass unimodal sentiment analysis.One challenge of multimodal sentiment analysis is how to design an efficient multimodal feature fusion strategy.Unfortunately,existing work always considers feature-level fusion or decision-level fusion,and few research works focus on hybrid fusion strategies that contain feature-level fusion and decision-level fusion.To improve the performance of multimodal sentiment analysis,we present a novel multimodal sentiment analysis model using BiGRU and attention-based hybrid fusion strategy(BAHFS).Firstly,we apply BiGRU to learn the unimodal features of text,audio and video.Then we fuse the unimodal features into bimodal features using the bimodal attention fusion module.Next,BAHFS feeds the unimodal features and bimodal features into the trimodal attention fusion module and the trimodal concatenation fusion module simultaneously to get two sets of trimodal features.Finally,BAHFS makes a classification with the two sets of trimodal features respectively and gets the final analysis results with decision-level fusion.Based on the CMU-MOSI and CMU-MOSEI datasets,extensive experiments have been carried out to verify BAHFS’s superiority. 展开更多
关键词 Multimdoal sentiment analysis BiGRU attention mechanism features-level fusion hybrid fusion strategy
下载PDF
Adaptive Multi-modal Fusion Instance Segmentation for CAEVs in Complex Conditions:Dataset,Framework and Verifications 被引量:1
7
作者 Pai Peng Keke Geng +3 位作者 Guodong Yin Yanbo Lu Weichao Zhuang Shuaipeng Liu 《Chinese Journal of Mechanical Engineering》 SCIE EI CAS CSCD 2021年第5期96-106,共11页
Current works of environmental perception for connected autonomous electrified vehicles(CAEVs)mainly focus on the object detection task in good weather and illumination conditions,they often perform poorly in adverse ... Current works of environmental perception for connected autonomous electrified vehicles(CAEVs)mainly focus on the object detection task in good weather and illumination conditions,they often perform poorly in adverse scenarios and have a vague scene parsing ability.This paper aims to develop an end-to-end sharpening mixture of experts(SMoE)fusion framework to improve the robustness and accuracy of the perception systems for CAEVs in complex illumination and weather conditions.Three original contributions make our work distinctive from the existing relevant literature.The Complex KITTI dataset is introduced which consists of 7481 pairs of modified KITTI RGB images and the generated LiDAR dense depth maps,and this dataset is fine annotated in instance-level with the proposed semi-automatic annotation method.The SMoE fusion approach is devised to adaptively learn the robust kernels from complementary modalities.Comprehensive comparative experiments are implemented,and the results show that the proposed SMoE framework yield significant improvements over the other fusion techniques in adverse environmental conditions.This research proposes a SMoE fusion framework to improve the scene parsing ability of the perception systems for CAEVs in adverse conditions. 展开更多
关键词 Connected autonomous electrified vehicles multi-modal fusion Semi-automatic annotation Sharpening mixture of experts Comparative experiments
下载PDF
Method of Multi-Mode Sensor Data Fusion with an Adaptive Deep Coupling Convolutional Auto-Encoder
8
作者 Xiaoxiong Feng Jianhua Liu 《Journal of Sensor Technology》 2023年第4期69-85,共17页
To address the difficulties in fusing multi-mode sensor data for complex industrial machinery, an adaptive deep coupling convolutional auto-encoder (ADCCAE) fusion method was proposed. First, the multi-mode features e... To address the difficulties in fusing multi-mode sensor data for complex industrial machinery, an adaptive deep coupling convolutional auto-encoder (ADCCAE) fusion method was proposed. First, the multi-mode features extracted synchronously by the CCAE were stacked and fed to the multi-channel convolution layers for fusion. Then, the fused data was passed to all connection layers for compression and fed to the Softmax module for classification. Finally, the coupling loss function coefficients and the network parameters were optimized through an adaptive approach using the gray wolf optimization (GWO) algorithm. Experimental comparisons showed that the proposed ADCCAE fusion model was superior to existing models for multi-mode data fusion. 展开更多
关键词 multi-mode Data fusion Coupling Convolutional Auto-Encoder Adaptive Optimization Deep Learning
下载PDF
Fake News Detection Based on Cross-Modal Message Aggregation and Gated Fusion Network
9
作者 Fangfang Shan Mengyao Liu +1 位作者 Menghan Zhang Zhenyu Wang 《Computers, Materials & Continua》 SCIE EI 2024年第7期1521-1542,共22页
Social media has become increasingly significant in modern society,but it has also turned into a breeding ground for the propagation of misleading information,potentially causing a detrimental impact on public opinion... Social media has become increasingly significant in modern society,but it has also turned into a breeding ground for the propagation of misleading information,potentially causing a detrimental impact on public opinion and daily life.Compared to pure text content,multmodal content significantly increases the visibility and share ability of posts.This has made the search for efficient modality representations and cross-modal information interaction methods a key focus in the field of multimodal fake news detection.To effectively address the critical challenge of accurately detecting fake news on social media,this paper proposes a fake news detection model based on crossmodal message aggregation and a gated fusion network(MAGF).MAGF first uses BERT to extract cumulative textual feature representations and word-level features,applies Faster Region-based ConvolutionalNeuralNetwork(Faster R-CNN)to obtain image objects,and leverages ResNet-50 and Visual Geometry Group-19(VGG-19)to obtain image region features and global features.The image region features and word-level text features are then projected into a low-dimensional space to calculate a text-image affinity matrix for cross-modal message aggregation.The gated fusion network combines text and image region features to obtain adaptively aggregated features.The interaction matrix is derived through an attention mechanism and further integrated with global image features using a co-attention mechanism to producemultimodal representations.Finally,these fused features are fed into a classifier for news categorization.Experiments were conducted on two public datasets,Twitter and Weibo.Results show that the proposed model achieves accuracy rates of 91.8%and 88.7%on the two datasets,respectively,significantly outperforming traditional unimodal and existing multimodal models. 展开更多
关键词 Fake news detection cross-modalmessage aggregation gate fusion network co-attention mechanism multi-modal representation
下载PDF
Fake News Detection Based on Text-Modal Dominance and Fusing Multiple Multi-Model Clues
10
作者 Li fang Fu Huanxin Peng +1 位作者 Changjin Ma Yuhan Liu 《Computers, Materials & Continua》 SCIE EI 2024年第3期4399-4416,共18页
In recent years,how to efficiently and accurately identify multi-model fake news has become more challenging.First,multi-model data provides more evidence but not all are equally important.Secondly,social structure in... In recent years,how to efficiently and accurately identify multi-model fake news has become more challenging.First,multi-model data provides more evidence but not all are equally important.Secondly,social structure information has proven to be effective in fake news detection and how to combine it while reducing the noise information is critical.Unfortunately,existing approaches fail to handle these problems.This paper proposes a multi-model fake news detection framework based on Tex-modal Dominance and fusing Multiple Multi-model Cues(TD-MMC),which utilizes three valuable multi-model clues:text-model importance,text-image complementary,and text-image inconsistency.TD-MMC is dominated by textural content and assisted by image information while using social network information to enhance text representation.To reduce the irrelevant social structure’s information interference,we use a unidirectional cross-modal attention mechanism to selectively learn the social structure’s features.A cross-modal attention mechanism is adopted to obtain text-image cross-modal features while retaining textual features to reduce the loss of important information.In addition,TD-MMC employs a new multi-model loss to improve the model’s generalization ability.Extensive experiments have been conducted on two public real-world English and Chinese datasets,and the results show that our proposed model outperforms the state-of-the-art methods on classification evaluation metrics. 展开更多
关键词 Fake news detection cross-modal attention mechanism multi-modal fusion social network transfer learning
下载PDF
Test method of laser paint removal based on multi-modal feature fusion
11
作者 HUANG Hai-peng HAO Ben-tian +2 位作者 YE De-jun GAO Hao LI Liang 《Journal of Central South University》 SCIE EI CAS CSCD 2022年第10期3385-3398,共14页
Laser cleaning is a highly nonlinear physical process for solving poor single-modal(e.g., acoustic or vision)detection performance and low inter-information utilization. In this study, a multi-modal feature fusion net... Laser cleaning is a highly nonlinear physical process for solving poor single-modal(e.g., acoustic or vision)detection performance and low inter-information utilization. In this study, a multi-modal feature fusion network model was constructed based on a laser paint removal experiment. The alignment of heterogeneous data under different modals was solved by combining the piecewise aggregate approximation and gramian angular field. Moreover, the attention mechanism was introduced to optimize the dual-path network and dense connection network, enabling the sampling characteristics to be extracted and integrated. Consequently, the multi-modal discriminant detection of laser paint removal was realized. According to the experimental results, the verification accuracy of the constructed model on the experimental dataset was 99.17%, which is 5.77% higher than the optimal single-modal detection results of the laser paint removal. The feature extraction network was optimized by the attention mechanism, and the model accuracy was increased by 3.3%. Results verify the improved classification performance of the constructed multi-modal feature fusion model in detecting laser paint removal, the effective integration of acoustic data and visual image data, and the accurate detection of laser paint removal. 展开更多
关键词 laser cleaning multi-modal fusion image processing deep learning
下载PDF
Adaptive multi-modal feature fusion for far and hard object detection
12
作者 LI Yang GE Hongwei 《Journal of Measurement Science and Instrumentation》 CAS CSCD 2021年第2期232-241,共10页
In order to solve difficult detection of far and hard objects due to the sparseness and insufficient semantic information of LiDAR point cloud,a 3D object detection network with multi-modal data adaptive fusion is pro... In order to solve difficult detection of far and hard objects due to the sparseness and insufficient semantic information of LiDAR point cloud,a 3D object detection network with multi-modal data adaptive fusion is proposed,which makes use of multi-neighborhood information of voxel and image information.Firstly,design an improved ResNet that maintains the structure information of far and hard objects in low-resolution feature maps,which is more suitable for detection task.Meanwhile,semantema of each image feature map is enhanced by semantic information from all subsequent feature maps.Secondly,extract multi-neighborhood context information with different receptive field sizes to make up for the defect of sparseness of point cloud which improves the ability of voxel features to represent the spatial structure and semantic information of objects.Finally,propose a multi-modal feature adaptive fusion strategy which uses learnable weights to express the contribution of different modal features to the detection task,and voxel attention further enhances the fused feature expression of effective target objects.The experimental results on the KITTI benchmark show that this method outperforms VoxelNet with remarkable margins,i.e.increasing the AP by 8.78%and 5.49%on medium and hard difficulty levels.Meanwhile,our method achieves greater detection performance compared with many mainstream multi-modal methods,i.e.outperforming the AP by 1%compared with that of MVX-Net on medium and hard difficulty levels. 展开更多
关键词 3D object detection adaptive fusion multi-modal data fusion attention mechanism multi-neighborhood features
下载PDF
Adaptive cross-fusion learning for multi-modal gesture recognition
13
作者 Benjia ZHOU Jun WAN +1 位作者 Yanyan LIANG Guodong GUO 《Virtual Reality & Intelligent Hardware》 2021年第3期235-247,共13页
Background Gesture recognition has attracted significant attention because of its wide range of potential applications.Although multi-modal gesture recognition has made significant progress in recent years,a popular m... Background Gesture recognition has attracted significant attention because of its wide range of potential applications.Although multi-modal gesture recognition has made significant progress in recent years,a popular method still is simply fusing prediction scores at the end of each branch,which often ignores complementary features among different modalities in the early stage and does not fuse the complementary features into a more discriminative feature.Methods This paper proposes an Adaptive Cross-modal Weighting(ACmW)scheme to exploit complementarity features from RGB-D data in this study.The scheme learns relations among different modalities by combining the features of different data streams.The proposed ACmW module contains two key functions:(1)fusing complementary features from multiple streams through an adaptive one-dimensional convolution;and(2)modeling the correlation of multi-stream complementary features in the time dimension.Through the effective combination of these two functional modules,the proposed ACmW can automatically analyze the relationship between the complementary features from different streams,and can fuse them in the spatial and temporal dimensions.Results Extensive experiments validate the effectiveness of the proposed method,and show that our method outperforms state-of-the-art methods on IsoGD and NVGesture. 展开更多
关键词 Gesture recognition multi-modal fusion RGB-D
下载PDF
Misuse of Teaching Gesture from the Perspective of Multi-modal Discourse Analysis and Its Solutions
14
作者 韩芳 《海外英语》 2018年第3期204-205,207,共3页
Based on the teaching video of middle school English teachers, through observation and analysis, it puts forward the problem of less use, wrong use and abuse in the use of teachers' teaching gestures in middle sch... Based on the teaching video of middle school English teachers, through observation and analysis, it puts forward the problem of less use, wrong use and abuse in the use of teachers' teaching gestures in middle school English teaching. And then it puts forward corresponding solutions from three aspects: concept, theory and practice. Hoping to provide further reference to the complementary role of teaching gesture and teaching discourse. 展开更多
关键词 multi-modal discourse teaching gesture MISUSE solution strategy
下载PDF
An intelligent navigation experimental system based on multi-mode fusion
15
作者 Rui HAN Zhiquan FENG +3 位作者 Jinglan TIAN Xue FAN Xiaohui YANG Qingbei GUO 《Virtual Reality & Intelligent Hardware》 2020年第4期345-353,共9页
At present,most experimental teaching systems lack guidance of an operator,and thus users often do not know what to do during an experiment.The user load is therefore increased,and the learning efficiency of the stude... At present,most experimental teaching systems lack guidance of an operator,and thus users often do not know what to do during an experiment.The user load is therefore increased,and the learning efficiency of the students is decreased.To solve the problem of insufficient system interactivity and guidance,an experimental navigation system based on multi-mode fusion is proposed in this paper.The system first obtains user information by sensing the hardware devices,intelligently perceives the user intention and progress of the experiment according to the information acquired,and finally carries out a multi-modal intelligent navigation process for users.As an innovative aspect of this study,an intelligent multi-mode navigation system is used to guide users in conducting experiments,thereby reducing the user load and enabling the users to effectively complete their experiments.The results prove that this system can guide users in completing their experiments,and can effectively reduce the user load during the interaction process and improve the efficiency. 展开更多
关键词 Navigation interaction Chemical experiment system multi-mode fusion
下载PDF
域迁移增强的综合假脸检测模型
16
作者 林新棋 董琳 +3 位作者 叶锋 肖觉斯 黄添强 黄丽清 《福建师范大学学报(自然科学版)》 CAS 北大核心 2024年第5期17-29,共13页
提出一种跨域的综合假脸检测模型。首先,设计一种双域融合模型,该模型利用空间注意力机制实现RGB域和频域特征的融合。其次,在此基础上,结合数据增强技术,提出了一种跨域迁移策略。最后,提出的双域模型的精度,在5个通用数据集上均比单... 提出一种跨域的综合假脸检测模型。首先,设计一种双域融合模型,该模型利用空间注意力机制实现RGB域和频域特征的融合。其次,在此基础上,结合数据增强技术,提出了一种跨域迁移策略。最后,提出的双域模型的精度,在5个通用数据集上均比单域模型有一定的提高,尤其在NT数据集上,该方法的精度比EfficientNet-B0方法提高了3.4%。此外,实验结果表明,与其他迁移学习方法相比,在FaceForensics++和Celeb-df数据集上,该方法在域迁移中具有更好的泛化性能。 展开更多
关键词 假脸检测 泛化能力 双域融合模型 迁移策略
下载PDF
多间隔信息融合的母线保护电流互感器断线再开放策略
17
作者 陈琦 陈福锋 +4 位作者 唐治国 薛明军 刘凯祥 王胜 孙震宇 《电力自动化设备》 EI CSCD 北大核心 2024年第5期151-157,共7页
对于电流互感器(CT)断线后发生金属性故障的情景,现有母线保护采取的闭锁差动保护动作方式将会引发多个变电站停电。此外,当母线区内发生高阻接地故障时,现有母线保护判据可能会将其误判为CT断线故障,不利于电力系统的安全稳定运行。针... 对于电流互感器(CT)断线后发生金属性故障的情景,现有母线保护采取的闭锁差动保护动作方式将会引发多个变电站停电。此外,当母线区内发生高阻接地故障时,现有母线保护判据可能会将其误判为CT断线故障,不利于电力系统的安全稳定运行。针对这一问题,提出了一种母线保护CT断线再开放策略,该策略基于断线间隔与非断线间隔的零序电流变化量对负荷波动和故障进行区分,并基于非断线间隔的差流有效值和间隔失灵保护信息对区内外故障进行判别。实时数字仿真系统仿真结果验证了所提CT断线识别判据和再开放策略的正确性。 展开更多
关键词 电流互感器断线 多间隔信息融合 短路故障 电流互感器闭锁策略 母线保护
下载PDF
基于融媒体的档案文化资源传播策略研究
18
作者 孟晓辉 《山西档案》 北大核心 2024年第4期130-133,共4页
随着融媒体时代的到来,媒介形态加速演进,信息传播生态发生深刻变革。作为历史文化遗产的重要组成部分,档案文化资源亟需借力媒介融合开启传播新局面。着眼媒介融合发展大势,围绕技术融合、内容融合、平台融合、用户融合4个维度,阐释了... 随着融媒体时代的到来,媒介形态加速演进,信息传播生态发生深刻变革。作为历史文化遗产的重要组成部分,档案文化资源亟需借力媒介融合开启传播新局面。着眼媒介融合发展大势,围绕技术融合、内容融合、平台融合、用户融合4个维度,阐释了档案文化资源传播的价值意义,对档案文化资源传播在理念滞后、渠道单一、内容同质、互动不足等方面的问题进行了深度剖析,从更新理念、优化渠道、创新内容、加强互动等层面系统探析了档案文化资源融媒体传播的策略路径,以期为新时代档案文化建设提供学理支撑和实践指引。 展开更多
关键词 档案文化资源 档案文化资源传播 融媒体
下载PDF
双通道深度图像先验降噪模型
19
作者 徐少平 肖楠 +2 位作者 罗洁 程晓慧 陈晓军 《电子学报》 EI CAS CSCD 北大核心 2024年第1期58-68,共11页
相对于采用固定网络参数值的有监督深度降噪模型而言,无监督的深度图像先验(Deep Image Prior,DIP)降噪模型更具灵活性和实用性.然而,DIP模型的降噪效果远低于有监督降噪模型(尤其是在处理人工合成噪声图像时).为进一步提升DIP降噪模型... 相对于采用固定网络参数值的有监督深度降噪模型而言,无监督的深度图像先验(Deep Image Prior,DIP)降噪模型更具灵活性和实用性.然而,DIP模型的降噪效果远低于有监督降噪模型(尤其是在处理人工合成噪声图像时).为进一步提升DIP降噪模型的降噪效果,本文提出了双通道深度图像先验降噪模型.该降噪模型由噪声图像预处理、在线迭代训练和图像融合3个模块组成.首先,利用BM3D和FFDNet两种经典降噪方法对给定的噪声图像进行预处理,得到2张初步降噪图像,然后,将原DIP单通道逼近目标图像架构拓展为双通道工作模式.其中,第一通道以FFDNet初步降噪图像和噪声图像为双目标图像,第二通道则以BM3D预处理图像和噪声图像为双目标图像.在此基础上,按照标准的DIP在线训练方式让DIP网络输出图像在两个通道上分别逼近各自的目标图像,同时依据基于边缘能量定义的伪有参考图像质量评价值适时终止迭代过程,从而获得2张中间生成图像.最后,使用结构化图块分解融合算法将两张中间生成图像融合并作为最终的降噪后图像.实验数据表明,在合成噪声图像上,本文提出的双通道深度图像先验降噪模型在各个噪声水平上显著优于原DIP及其他无监督降噪模型(提升了约2.2 dB),甚至逼近和超过了新近提出的主流有监督降噪模型,这充分表明了本文提出的改进策略的有效性;在真实噪声图像上,本文提出的降噪模型优于排名第二的对比降噪方法约2 dB,展现出其在实际应用场景下独有的优势. 展开更多
关键词 深度图像先验 双通道逼近策略 预处理图像 自动迭代终止 图像质量评价 图像融合
下载PDF
视觉融合下的复杂路况车辆换道决策模型
20
作者 袁健 陈佳钦 +2 位作者 潘杰忠 孙煜 赵逢禹 《小型微型计算机系统》 CSCD 北大核心 2024年第9期2205-2214,共10页
车辆换道是一个融合车辆行为及其周围环境的多因素动态过程,不合理的换道行为会招致严重的交通事故.现有安全防撞模型未考虑换道时多车辆间的耦合关系,且换道车辆的感知方式有一定局限性,在需强制性换道时决策效果不理想.因此本文提出... 车辆换道是一个融合车辆行为及其周围环境的多因素动态过程,不合理的换道行为会招致严重的交通事故.现有安全防撞模型未考虑换道时多车辆间的耦合关系,且换道车辆的感知方式有一定局限性,在需强制性换道时决策效果不理想.因此本文提出视觉融合下的复杂路况车辆换道决策模型.该模型以单视角补偿双目视差计算的视觉融合方法感知参与换道的多车辆,解决了因车辆遮挡覆盖等非线性运动导致的多目标车辆轨迹预测误差问题,通过提取出三维的换道行为参数输入RBF神经网络进行可行性安全评估,可使车辆在合适时机进行换道.实验结果表明,提出的车辆换道决策模型在相同实验环境下相较单目检测算法YOLOv3准确性可提升5.1%,相较选取的基准双目检测算法准确性提升0.7%,且对换道进行安全评估的预测综合准确率达97.33%,能满足自动驾驶车辆强制性换道需求. 展开更多
关键词 换道决策 自动驾驶车辆 视觉融合策略 换道安全评估 强制性换道
下载PDF
上一页 1 2 25 下一页 到第
使用帮助 返回顶部