In recent years,how to efficiently and accurately identify multi-model fake news has become more challenging.First,multi-model data provides more evidence but not all are equally important.Secondly,social structure in...In recent years,how to efficiently and accurately identify multi-model fake news has become more challenging.First,multi-model data provides more evidence but not all are equally important.Secondly,social structure information has proven to be effective in fake news detection and how to combine it while reducing the noise information is critical.Unfortunately,existing approaches fail to handle these problems.This paper proposes a multi-model fake news detection framework based on Tex-modal Dominance and fusing Multiple Multi-model Cues(TD-MMC),which utilizes three valuable multi-model clues:text-model importance,text-image complementary,and text-image inconsistency.TD-MMC is dominated by textural content and assisted by image information while using social network information to enhance text representation.To reduce the irrelevant social structure’s information interference,we use a unidirectional cross-modal attention mechanism to selectively learn the social structure’s features.A cross-modal attention mechanism is adopted to obtain text-image cross-modal features while retaining textual features to reduce the loss of important information.In addition,TD-MMC employs a new multi-model loss to improve the model’s generalization ability.Extensive experiments have been conducted on two public real-world English and Chinese datasets,and the results show that our proposed model outperforms the state-of-the-art methods on classification evaluation metrics.展开更多
In mobile machinery,hydro-mechanical pumps are increasingly replaced by electronically controlled pumps to improve the automation level,but diversified control functions(e.g.,power limitation and pressure cut-off)are ...In mobile machinery,hydro-mechanical pumps are increasingly replaced by electronically controlled pumps to improve the automation level,but diversified control functions(e.g.,power limitation and pressure cut-off)are integrated into the electronic controller only from the pump level,leading to the potential instability of the overall system.To solve this problem,a multi-mode electrohydraulic load sensing(MELS)control scheme is proposed especially considering the switching stability from the system level,which includes four working modes of flow control,load sensing,power limitation,and pressure control.Depending on the actual working requirements,the switching rules for the different modes and the switching direction(i.e.,the modes can be switched bilaterally or unilaterally)are defined.The priority of different modes is also defined,from high to low:pressure control,power limitation,load sensing,and flow control.When multiple switching rules are satisfied at the same time,the system switches to the control mode with the highest priority.In addition,the switching stability between flow control and pressure control modes is analyzed,and the controller parameters that guarantee the switching stability are obtained.A comparative study is carried out based on a test rig with a 2-ton hydraulic excavator.The results show that the MELS controller can achieve the control functions of proper flow supplement,power limitation,and pressure cut-off,which has good stability performance when switching between different control modes.This research proposes the MELS control method that realizes the stability of multi-mode switching of the hydraulic system of mobile machinery under different working conditions.展开更多
With the development of Global Navigation Satellite Systems(GNSS),geodetic GNSS receivers have been utilized to monitor sea levels using GNSS-Interferometry Reflectometry(GNSS-IR)technology.The multi-mode,multi-freque...With the development of Global Navigation Satellite Systems(GNSS),geodetic GNSS receivers have been utilized to monitor sea levels using GNSS-Interferometry Reflectometry(GNSS-IR)technology.The multi-mode,multi-frequency signals of GPS,GLONASS,Galileo,and Beidou can be used for GNSS-IR sea level retrieval,but combining these retrievals remains problematic.To address this issue,a GNSS-IR sea level retrieval combination system has been developed,which begins by analyzing error sources in GNSS-IR sea level retrieval and establishing and solving the GNSS-IR retrieval equation.This paper focuses on two key points:time window selection and equation stability.The stability of the retrieval combination equations is determined by the condition number of the coefficient matrix within the time window.The impact of ill-conditioned coefficient matrices on the retrieval results is demonstrated using an extreme case of SNR data with only ascending or descending trajectories.After determining the time window and removing ill-conditioned equations,the multi-mode,multi-frequency GNSS-IR retrieval is performed.Results from three International GNSS Service(IGS)stations show that the combination method produces high-precision,high-resolution,and high-reliability sea level retrieval combination sequences.展开更多
To address the difficulties in fusing multi-mode sensor data for complex industrial machinery, an adaptive deep coupling convolutional auto-encoder (ADCCAE) fusion method was proposed. First, the multi-mode features e...To address the difficulties in fusing multi-mode sensor data for complex industrial machinery, an adaptive deep coupling convolutional auto-encoder (ADCCAE) fusion method was proposed. First, the multi-mode features extracted synchronously by the CCAE were stacked and fed to the multi-channel convolution layers for fusion. Then, the fused data was passed to all connection layers for compression and fed to the Softmax module for classification. Finally, the coupling loss function coefficients and the network parameters were optimized through an adaptive approach using the gray wolf optimization (GWO) algorithm. Experimental comparisons showed that the proposed ADCCAE fusion model was superior to existing models for multi-mode data fusion.展开更多
Fusing hand-based features in multi-modal biometric recognition enhances anti-spoofing capabilities.Additionally,it leverages inter-modal correlation to enhance recognition performance.Concurrently,the robustness and ...Fusing hand-based features in multi-modal biometric recognition enhances anti-spoofing capabilities.Additionally,it leverages inter-modal correlation to enhance recognition performance.Concurrently,the robustness and recognition performance of the system can be enhanced through judiciously leveraging the correlation among multimodal features.Nevertheless,two issues persist in multi-modal feature fusion recognition:Firstly,the enhancement of recognition performance in fusion recognition has not comprehensively considered the inter-modality correlations among distinct modalities.Secondly,during modal fusion,improper weight selection diminishes the salience of crucial modal features,thereby diminishing the overall recognition performance.To address these two issues,we introduce an enhanced DenseNet multimodal recognition network founded on feature-level fusion.The information from the three modalities is fused akin to RGB,and the input network augments the correlation between modes through channel correlation.Within the enhanced DenseNet network,the Efficient Channel Attention Network(ECA-Net)dynamically adjusts the weight of each channel to amplify the salience of crucial information in each modal feature.Depthwise separable convolution markedly reduces the training parameters and further enhances the feature correlation.Experimental evaluations were conducted on four multimodal databases,comprising six unimodal databases,including multispectral palmprint and palm vein databases from the Chinese Academy of Sciences.The Equal Error Rates(EER)values were 0.0149%,0.0150%,0.0099%,and 0.0050%,correspondingly.In comparison to other network methods for palmprint,palm vein,and finger vein fusion recognition,this approach substantially enhances recognition performance,rendering it suitable for high-security environments with practical applicability.The experiments in this article utilized amodest sample database comprising 200 individuals.The subsequent phase involves preparing for the extension of the method to larger databases.展开更多
Media convergence works by processing information from different modalities and applying them to different domains.It is difficult for the conventional knowledge graph to utilise multi-media features because the intro...Media convergence works by processing information from different modalities and applying them to different domains.It is difficult for the conventional knowledge graph to utilise multi-media features because the introduction of a large amount of information from other modalities reduces the effectiveness of representation learning and makes knowledge graph inference less effective.To address the issue,an inference method based on Media Convergence and Rule-guided Joint Inference model(MCRJI)has been pro-posed.The authors not only converge multi-media features of entities but also introduce logic rules to improve the accuracy and interpretability of link prediction.First,a multi-headed self-attention approach is used to obtain the attention of different media features of entities during semantic synthesis.Second,logic rules of different lengths are mined from knowledge graph to learn new entity representations.Finally,knowledge graph inference is performed based on representing entities that converge multi-media features.Numerous experimental results show that MCRJI outperforms other advanced baselines in using multi-media features and knowledge graph inference,demonstrating that MCRJI provides an excellent approach for knowledge graph inference with converged multi-media features.展开更多
Recently,there have been significant advancements in the study of semantic communication in single-modal scenarios.However,the ability to process information in multi-modal environments remains limited.Inspired by the...Recently,there have been significant advancements in the study of semantic communication in single-modal scenarios.However,the ability to process information in multi-modal environments remains limited.Inspired by the research and applications of natural language processing across different modalities,our goal is to accurately extract frame-level semantic information from videos and ultimately transmit high-quality videos.Specifically,we propose a deep learning-basedMulti-ModalMutual Enhancement Video Semantic Communication system,called M3E-VSC.Built upon a VectorQuantized Generative AdversarialNetwork(VQGAN),our systemaims to leverage mutual enhancement among different modalities by using text as the main carrier of transmission.With it,the semantic information can be extracted fromkey-frame images and audio of the video and performdifferential value to ensure that the extracted text conveys accurate semantic information with fewer bits,thus improving the capacity of the system.Furthermore,a multi-frame semantic detection module is designed to facilitate semantic transitions during video generation.Simulation results demonstrate that our proposed model maintains high robustness in complex noise environments,particularly in low signal-to-noise ratio conditions,significantly improving the accuracy and speed of semantic transmission in video communication by approximately 50 percent.展开更多
The unsupervised multi-modal image translation is an emerging domain of computer vision whose goal is to transform an image from the source domain into many diverse styles in the target domain.However,the multi-genera...The unsupervised multi-modal image translation is an emerging domain of computer vision whose goal is to transform an image from the source domain into many diverse styles in the target domain.However,the multi-generator mechanism is employed among the advanced approaches available to model different domain mappings,which results in inefficient training of neural networks and pattern collapse,leading to inefficient generation of image diversity.To address this issue,this paper introduces a multi-modal unsupervised image translation framework that uses a generator to perform multi-modal image translation.Specifically,firstly,the domain code is introduced in this paper to explicitly control the different generation tasks.Secondly,this paper brings in the squeeze-and-excitation(SE)mechanism and feature attention(FA)module.Finally,the model integrates multiple optimization objectives to ensure efficient multi-modal translation.This paper performs qualitative and quantitative experiments on multiple non-paired benchmark image translation datasets while demonstrating the benefits of the proposed method over existing technologies.Overall,experimental results have shown that the proposed method is versatile and scalable.展开更多
For the multi-mode radar working in the modern electronicbattlefield, different working states of one single radar areprone to being classified as multiple emitters when adoptingtraditional classification methods to p...For the multi-mode radar working in the modern electronicbattlefield, different working states of one single radar areprone to being classified as multiple emitters when adoptingtraditional classification methods to process intercepted signals,which has a negative effect on signal classification. A classificationmethod based on spatial data mining is presented to address theabove challenge. Inspired by the idea of spatial data mining, theclassification method applies nuclear field to depicting the distributioninformation of pulse samples in feature space, and digs out thehidden cluster information by analyzing distribution characteristics.In addition, a membership-degree criterion to quantify the correlationamong all classes is established, which ensures classificationaccuracy of signal samples. Numerical experiments show that thepresented method can effectively prevent different working statesof multi-mode emitter from being classified as several emitters,and achieve higher classification accuracy.展开更多
There are multiple operating modes in the real industrial process, and the collected data follow the complex multimodal distribution, so most traditional process monitoring methods are no longer applicable because the...There are multiple operating modes in the real industrial process, and the collected data follow the complex multimodal distribution, so most traditional process monitoring methods are no longer applicable because their presumptions are that sampled-data should obey the single Gaussian distribution or non-Gaussian distribution. In order to solve these problems, a novel weighted local standardization(WLS) strategy is proposed to standardize the multimodal data, which can eliminate the multi-mode characteristics of the collected data, and normalize them into unimodal data distribution. After detailed analysis of the raised data preprocessing strategy, a new algorithm using WLS strategy with support vector data description(SVDD) is put forward to apply for multi-mode monitoring process. Unlike the strategy of building multiple local models, the developed method only contains a model without the prior knowledge of multi-mode process. To demonstrate the proposed method's validity, it is applied to a numerical example and a Tennessee Eastman(TE) process. Finally, the simulation results show that the WLS strategy is very effective to standardize multimodal data, and the WLS-SVDD monitoring method has great advantages over the traditional SVDD and PCA combined with a local standardization strategy(LNS-PCA) in multi-mode process monitoring.展开更多
Based on the analysis of B3G evolution, the base-band processing chips for mobile terminals are introduced. Key technologies for multi-mode mobile terminal base-band chips are discussed. Terminal technologies are thou...Based on the analysis of B3G evolution, the base-band processing chips for mobile terminals are introduced. Key technologies for multi-mode mobile terminal base-band chips are discussed. Terminal technologies are thought to be the key of B3G, and terminal base-band chips are regarded as the core of terminal technologies. Therefore, a unified wireless development platform is required for the R&D of multi-mode mobile terminal base-band processing chips.展开更多
Multi-mode power internet of things(PIoT)combines various communication media to provide spatio-temporal coverage for low-carbon operation in smart park.Edge-end collaboration is feasible to achieve the full utilizati...Multi-mode power internet of things(PIoT)combines various communication media to provide spatio-temporal coverage for low-carbon operation in smart park.Edge-end collaboration is feasible to achieve the full utilization of heterogeneous resources and anti-eavesdropping.However,edge-end collaboration-based multi-mode PIoT faces challenges of mutual contradiction in communication and security quality of service(QoS)guarantee,inadaptability of resource management,and multi-mode access conflict.We propose an Adaptive learning based delAysensitive and seCure Edge-End Collaboration algorithm(ACE_(2))to optimize multi-mode channel selection and split device power into artificial noise(AN)transmission and data transmission for secure data delivery.ACE_(2) can achieve multi-attribute QoS guarantee,adaptive resource management and security enhancement,and access conflict elimination with the combined power of deep actor-critic(DAC),“win or learn fast(WoLF)”mechanism,and edge-end collaboration.Simulations demonstrate its superior performance in queuing delay,energy consumption,secrecy capacity,and adaptability to differentiated low-carbon services.展开更多
This paper studies the application of mathematical models to analyze the vortex-induced vibrations of the tendons of a given TLP along the Indian coastline, by using an analytical approach, using MATLAB. The tendon is...This paper studies the application of mathematical models to analyze the vortex-induced vibrations of the tendons of a given TLP along the Indian coastline, by using an analytical approach, using MATLAB. The tendon is subjected to a steady current load, which causes vortex-shedding downstream, leading to cross-flow vibrations. The magnitude of the excitation(lift and drag coefficients) depends on the vortex-shedding frequency. The resulting vibration is studied for possible resonant behavior. The excitation force is quantified empirically, the added mass by potential flow hydrodynamics, and the vibration by normal mode summation method. Non-linear viscous damping of the water is considered. The non-linear oscillations are studied by the phase-plane method, investigating the limit-cycle oscillations. The stable/unstable regions of the dynamic behavior are demarcated. The modal contribution to the total deflection is studied to establish the possibility of resonance of one of the wet modes with the vortex-shedding frequency.展开更多
The variability of the air-sea system in the low-frequency time domain can be decomposed into several systematic climate modes, namely, the decadal variability (DV) mode, the El Nino Southem Oscillation (ENSO) mod...The variability of the air-sea system in the low-frequency time domain can be decomposed into several systematic climate modes, namely, the decadal variability (DV) mode, the El Nino Southem Oscillation (ENSO) mode, the annual cycle (AC) mode, the semiannual cycle ( SC ) mode and the intraseasonal variability ( ISV ) mode. The combination of these primary modes in the air - sea system orchestrates a complex climate system. The multi-mode low-frequency variability in SST is investigated based on 22 a SST records from 1982 through 2003. The variation of SST in the past two decades undergoes a different combination of these dominant climate modes over different regions, which leads to an interesting new classification of the global ocean based on the relative importance of these modes. The new classification can provide ideal locations for better monitoring of these low-frequency modes in the scientific proof sense. Moreover, two no-annual variation and 14 no-semiannual variation oceanic points, termed annual and semiannual amphidromes, have been well defined in the AC and SC phase maps. The formation of these nodal points is attributed to the couplings of climate modes in EOF analysis results.展开更多
Probability-based diagnostic imaging(PDI)is one of the most well-known damage identification methods using guided waves.It is usually applied to diagnose damage in plates.The previous studies were dependent on the cer...Probability-based diagnostic imaging(PDI)is one of the most well-known damage identification methods using guided waves.It is usually applied to diagnose damage in plates.The previous studies were dependent on the certain damage index(DI)which is always calculated from the guided wave signals.In conventional methods,DI is simply defined by comparing the real-time data with the baseline data as reference.However,the baseline signal is easily affected by varying environmental conditions of structures.In this paper,a reference-free diagnostic imaging method is developed to avoid the influence of environmental factors,such as temperature and load conditions.The DI is defined based on the mode conversion of multi-mode guided waves with realtime signals without baseline signals.To improve the accuracy of diagnosis,two terms are included in the reference-free DI.One is called energy DI,which is defined based on the feature of signal energy.The other is called correlation DI and is defined based on the correlation coefficient.Then the PDI algorithm can be carried out instantaneously according to the reference-free DI.The real-time signals which are used to calculate DI are collected by the piezoelectric lead zirconate titanate(PZT)transducers placed on both sides of a plate.The numerical simulations by the finite element(FE)method on aluminum plates with PZT arrays are performed to validate the effectiveness of the reference-free damage diagnostic imaging.The approach is validated by two different arrays:a circle network and a square network.The results of diagnostic imaging are demonstrated and discussed in this paper.Furthermore,the advantage of reference-free DI is investigated by comparing the accuracy of defined reference-free DI and energy DI.展开更多
This paper presents a reconfigurable RF front-end for multi-mode multi-standard(MMMS) applications. The designed RF front-end is fabricated in 0.18 μm RF CMOS technology. The low noise characteristic is achieved by t...This paper presents a reconfigurable RF front-end for multi-mode multi-standard(MMMS) applications. The designed RF front-end is fabricated in 0.18 μm RF CMOS technology. The low noise characteristic is achieved by the noise canceling technique while the bandwidth is enhanced by gate inductive peaking technique. Measurement results show that, while the input frequency ranges from 100 MHz to 2.9 GHz, the proposed reconfigurable RF front-end achieves a controllable voltage conversion gain(VCG) from 18 dB to 39 dB. The measured maximum input third intercept point(IIP3) is-4.9 dBm and the minimum noise figure(NF) is 4.6 dB. The consumed current ranges from 16 mA to 26.5 mA from a 1.8 V supply voltage. The chip occupies an area of 1.17 mm^2 including pads.展开更多
A memetic algorithm (MA) for a multi-mode resourceconstrained project scheduling problem (MRCPSP) is proposed. We use a new fitness function and two very effective local search procedures in the proposed MA. The f...A memetic algorithm (MA) for a multi-mode resourceconstrained project scheduling problem (MRCPSP) is proposed. We use a new fitness function and two very effective local search procedures in the proposed MA. The fitness function makes use of a mechanism called "strategic oscillation" to make the search process have a higher probability to visit solutions around a "feasible boundary". One of the local search procedures aims at improving the lower bound of project makespan to be less than a known upper bound, and another aims at improving a solution of an MRCPSP instance accepting infeasible solutions based on the new fitness function in the search process. A detailed computational experiment is set up using instances from the problem instance library PSPLIB. Computational results show that the proposed MA is very competitive with the state-of-the-art algorithms. The MA obtains improved solutions for one instance of set J30.展开更多
The sixth generation(6G)of mobile communication system is witnessing a new paradigm shift,i.e.,integrated sensing-communication system.A comprehensive dataset is a prerequisite for 6G integrated sensing-communication ...The sixth generation(6G)of mobile communication system is witnessing a new paradigm shift,i.e.,integrated sensing-communication system.A comprehensive dataset is a prerequisite for 6G integrated sensing-communication research.This paper develops a novel simulation dataset,named M3SC,for mixed multi-modal(MMM)sensing-communication integration,and the generation framework of the M3SC dataset is further given.To obtain multimodal sensory data in physical space and communication data in electromagnetic space,we utilize Air-Sim and WaveFarer to collect multi-modal sensory data and exploit Wireless InSite to collect communication data.Furthermore,the in-depth integration and precise alignment of AirSim,WaveFarer,andWireless InSite are achieved.The M3SC dataset covers various weather conditions,multiplex frequency bands,and different times of the day.Currently,the M3SC dataset contains 1500 snapshots,including 80 RGB images,160 depth maps,80 LiDAR point clouds,256 sets of mmWave waveforms with 8 radar point clouds,and 72 channel impulse response(CIR)matrices per snapshot,thus totaling 120,000 RGB images,240,000 depth maps,120,000 LiDAR point clouds,384,000 sets of mmWave waveforms with 12,000 radar point clouds,and 108,000 CIR matrices.The data processing result presents the multi-modal sensory information and communication channel statistical properties.Finally,the MMM sensing-communication application,which can be supported by the M3SC dataset,is discussed.展开更多
Most visual privacy protection methods only hide the identity information of the face images,but the expression,behavior and some other information,which are of great significant in the live broadcast and other scenar...Most visual privacy protection methods only hide the identity information of the face images,but the expression,behavior and some other information,which are of great significant in the live broadcast and other scenarios,are also destroyed by the privacy protection process.To this end,this paper introduces a method to remove the identity information while preserving the expression information by performing multi-mode discriminant analysis on the images normalized with AAM algorithm.The face images are decomposed into mutually orthogonal subspaces corresponding to face attributes such as gender,race and expression,each of which owns related characteristic parameters.Then,the expression parameter is preserves to keep the facial expression information while others parameters,including gender and race,are modified to protect face privacy.The experiments show that this method yields well performance on both data utility and privacy protection.展开更多
基金This research was funded by the General Project of Philosophy and Social Science of Heilongjiang Province,Grant Number:20SHB080.
文摘In recent years,how to efficiently and accurately identify multi-model fake news has become more challenging.First,multi-model data provides more evidence but not all are equally important.Secondly,social structure information has proven to be effective in fake news detection and how to combine it while reducing the noise information is critical.Unfortunately,existing approaches fail to handle these problems.This paper proposes a multi-model fake news detection framework based on Tex-modal Dominance and fusing Multiple Multi-model Cues(TD-MMC),which utilizes three valuable multi-model clues:text-model importance,text-image complementary,and text-image inconsistency.TD-MMC is dominated by textural content and assisted by image information while using social network information to enhance text representation.To reduce the irrelevant social structure’s information interference,we use a unidirectional cross-modal attention mechanism to selectively learn the social structure’s features.A cross-modal attention mechanism is adopted to obtain text-image cross-modal features while retaining textual features to reduce the loss of important information.In addition,TD-MMC employs a new multi-model loss to improve the model’s generalization ability.Extensive experiments have been conducted on two public real-world English and Chinese datasets,and the results show that our proposed model outperforms the state-of-the-art methods on classification evaluation metrics.
基金National Key Research and Development Program of China(Grant No.2020YFB2009702)National Natural Science Foundation of China(Grant Nos.52075055,U21A20124 and 52111530069)Chongqing Natural Science Foundation of China(Grant No.cstc2020jcyj-msxmX0780)。
文摘In mobile machinery,hydro-mechanical pumps are increasingly replaced by electronically controlled pumps to improve the automation level,but diversified control functions(e.g.,power limitation and pressure cut-off)are integrated into the electronic controller only from the pump level,leading to the potential instability of the overall system.To solve this problem,a multi-mode electrohydraulic load sensing(MELS)control scheme is proposed especially considering the switching stability from the system level,which includes four working modes of flow control,load sensing,power limitation,and pressure control.Depending on the actual working requirements,the switching rules for the different modes and the switching direction(i.e.,the modes can be switched bilaterally or unilaterally)are defined.The priority of different modes is also defined,from high to low:pressure control,power limitation,load sensing,and flow control.When multiple switching rules are satisfied at the same time,the system switches to the control mode with the highest priority.In addition,the switching stability between flow control and pressure control modes is analyzed,and the controller parameters that guarantee the switching stability are obtained.A comparative study is carried out based on a test rig with a 2-ton hydraulic excavator.The results show that the MELS controller can achieve the control functions of proper flow supplement,power limitation,and pressure cut-off,which has good stability performance when switching between different control modes.This research proposes the MELS control method that realizes the stability of multi-mode switching of the hydraulic system of mobile machinery under different working conditions.
基金National Natural Science Foundation of China(No.42004018)。
文摘With the development of Global Navigation Satellite Systems(GNSS),geodetic GNSS receivers have been utilized to monitor sea levels using GNSS-Interferometry Reflectometry(GNSS-IR)technology.The multi-mode,multi-frequency signals of GPS,GLONASS,Galileo,and Beidou can be used for GNSS-IR sea level retrieval,but combining these retrievals remains problematic.To address this issue,a GNSS-IR sea level retrieval combination system has been developed,which begins by analyzing error sources in GNSS-IR sea level retrieval and establishing and solving the GNSS-IR retrieval equation.This paper focuses on two key points:time window selection and equation stability.The stability of the retrieval combination equations is determined by the condition number of the coefficient matrix within the time window.The impact of ill-conditioned coefficient matrices on the retrieval results is demonstrated using an extreme case of SNR data with only ascending or descending trajectories.After determining the time window and removing ill-conditioned equations,the multi-mode,multi-frequency GNSS-IR retrieval is performed.Results from three International GNSS Service(IGS)stations show that the combination method produces high-precision,high-resolution,and high-reliability sea level retrieval combination sequences.
文摘To address the difficulties in fusing multi-mode sensor data for complex industrial machinery, an adaptive deep coupling convolutional auto-encoder (ADCCAE) fusion method was proposed. First, the multi-mode features extracted synchronously by the CCAE were stacked and fed to the multi-channel convolution layers for fusion. Then, the fused data was passed to all connection layers for compression and fed to the Softmax module for classification. Finally, the coupling loss function coefficients and the network parameters were optimized through an adaptive approach using the gray wolf optimization (GWO) algorithm. Experimental comparisons showed that the proposed ADCCAE fusion model was superior to existing models for multi-mode data fusion.
基金funded by the National Natural Science Foundation of China(61991413)the China Postdoctoral Science Foundation(2019M651142)+1 种基金the Natural Science Foundation of Liaoning Province(2021-KF-12-07)the Natural Science Foundations of Liaoning Province(2023-MS-322).
文摘Fusing hand-based features in multi-modal biometric recognition enhances anti-spoofing capabilities.Additionally,it leverages inter-modal correlation to enhance recognition performance.Concurrently,the robustness and recognition performance of the system can be enhanced through judiciously leveraging the correlation among multimodal features.Nevertheless,two issues persist in multi-modal feature fusion recognition:Firstly,the enhancement of recognition performance in fusion recognition has not comprehensively considered the inter-modality correlations among distinct modalities.Secondly,during modal fusion,improper weight selection diminishes the salience of crucial modal features,thereby diminishing the overall recognition performance.To address these two issues,we introduce an enhanced DenseNet multimodal recognition network founded on feature-level fusion.The information from the three modalities is fused akin to RGB,and the input network augments the correlation between modes through channel correlation.Within the enhanced DenseNet network,the Efficient Channel Attention Network(ECA-Net)dynamically adjusts the weight of each channel to amplify the salience of crucial information in each modal feature.Depthwise separable convolution markedly reduces the training parameters and further enhances the feature correlation.Experimental evaluations were conducted on four multimodal databases,comprising six unimodal databases,including multispectral palmprint and palm vein databases from the Chinese Academy of Sciences.The Equal Error Rates(EER)values were 0.0149%,0.0150%,0.0099%,and 0.0050%,correspondingly.In comparison to other network methods for palmprint,palm vein,and finger vein fusion recognition,this approach substantially enhances recognition performance,rendering it suitable for high-security environments with practical applicability.The experiments in this article utilized amodest sample database comprising 200 individuals.The subsequent phase involves preparing for the extension of the method to larger databases.
基金National College Students’Training Programs of Innovation and Entrepreneurship,Grant/Award Number:S202210022060the CACMS Innovation Fund,Grant/Award Number:CI2021A00512the National Nature Science Foundation of China under Grant,Grant/Award Number:62206021。
文摘Media convergence works by processing information from different modalities and applying them to different domains.It is difficult for the conventional knowledge graph to utilise multi-media features because the introduction of a large amount of information from other modalities reduces the effectiveness of representation learning and makes knowledge graph inference less effective.To address the issue,an inference method based on Media Convergence and Rule-guided Joint Inference model(MCRJI)has been pro-posed.The authors not only converge multi-media features of entities but also introduce logic rules to improve the accuracy and interpretability of link prediction.First,a multi-headed self-attention approach is used to obtain the attention of different media features of entities during semantic synthesis.Second,logic rules of different lengths are mined from knowledge graph to learn new entity representations.Finally,knowledge graph inference is performed based on representing entities that converge multi-media features.Numerous experimental results show that MCRJI outperforms other advanced baselines in using multi-media features and knowledge graph inference,demonstrating that MCRJI provides an excellent approach for knowledge graph inference with converged multi-media features.
基金supported by the National Key Research and Development Project under Grant 2020YFB1807602Key Program of Marine Economy Development Special Foundation of Department of Natural Resources of Guangdong Province(GDNRC[2023]24)the National Natural Science Foundation of China under Grant 62271267.
文摘Recently,there have been significant advancements in the study of semantic communication in single-modal scenarios.However,the ability to process information in multi-modal environments remains limited.Inspired by the research and applications of natural language processing across different modalities,our goal is to accurately extract frame-level semantic information from videos and ultimately transmit high-quality videos.Specifically,we propose a deep learning-basedMulti-ModalMutual Enhancement Video Semantic Communication system,called M3E-VSC.Built upon a VectorQuantized Generative AdversarialNetwork(VQGAN),our systemaims to leverage mutual enhancement among different modalities by using text as the main carrier of transmission.With it,the semantic information can be extracted fromkey-frame images and audio of the video and performdifferential value to ensure that the extracted text conveys accurate semantic information with fewer bits,thus improving the capacity of the system.Furthermore,a multi-frame semantic detection module is designed to facilitate semantic transitions during video generation.Simulation results demonstrate that our proposed model maintains high robustness in complex noise environments,particularly in low signal-to-noise ratio conditions,significantly improving the accuracy and speed of semantic transmission in video communication by approximately 50 percent.
基金the National Natural Science Foundation of China(No.61976080)the Academic Degrees&Graduate Education Reform Project of Henan Province(No.2021SJGLX195Y)+1 种基金the Teaching Reform Research and Practice Project of Henan Undergraduate Universities(No.2022SYJXLX008)the Key Project on Research and Practice of Henan University Graduate Education and Teaching Reform(No.YJSJG2023XJ006)。
文摘The unsupervised multi-modal image translation is an emerging domain of computer vision whose goal is to transform an image from the source domain into many diverse styles in the target domain.However,the multi-generator mechanism is employed among the advanced approaches available to model different domain mappings,which results in inefficient training of neural networks and pattern collapse,leading to inefficient generation of image diversity.To address this issue,this paper introduces a multi-modal unsupervised image translation framework that uses a generator to perform multi-modal image translation.Specifically,firstly,the domain code is introduced in this paper to explicitly control the different generation tasks.Secondly,this paper brings in the squeeze-and-excitation(SE)mechanism and feature attention(FA)module.Finally,the model integrates multiple optimization objectives to ensure efficient multi-modal translation.This paper performs qualitative and quantitative experiments on multiple non-paired benchmark image translation datasets while demonstrating the benefits of the proposed method over existing technologies.Overall,experimental results have shown that the proposed method is versatile and scalable.
基金supported by the National Natural Science Foundation of China(61371172)the International S&T Cooperation Program of China(2015DFR10220)+1 种基金the Ocean Engineering Project of National Key Laboratory Foundation(1213)the Fundamental Research Funds for the Central Universities(HEUCF1608)
文摘For the multi-mode radar working in the modern electronicbattlefield, different working states of one single radar areprone to being classified as multiple emitters when adoptingtraditional classification methods to process intercepted signals,which has a negative effect on signal classification. A classificationmethod based on spatial data mining is presented to address theabove challenge. Inspired by the idea of spatial data mining, theclassification method applies nuclear field to depicting the distributioninformation of pulse samples in feature space, and digs out thehidden cluster information by analyzing distribution characteristics.In addition, a membership-degree criterion to quantify the correlationamong all classes is established, which ensures classificationaccuracy of signal samples. Numerical experiments show that thepresented method can effectively prevent different working statesof multi-mode emitter from being classified as several emitters,and achieve higher classification accuracy.
基金Project(61374140)supported by the National Natural Science Foundation of China
文摘There are multiple operating modes in the real industrial process, and the collected data follow the complex multimodal distribution, so most traditional process monitoring methods are no longer applicable because their presumptions are that sampled-data should obey the single Gaussian distribution or non-Gaussian distribution. In order to solve these problems, a novel weighted local standardization(WLS) strategy is proposed to standardize the multimodal data, which can eliminate the multi-mode characteristics of the collected data, and normalize them into unimodal data distribution. After detailed analysis of the raised data preprocessing strategy, a new algorithm using WLS strategy with support vector data description(SVDD) is put forward to apply for multi-mode monitoring process. Unlike the strategy of building multiple local models, the developed method only contains a model without the prior knowledge of multi-mode process. To demonstrate the proposed method's validity, it is applied to a numerical example and a Tennessee Eastman(TE) process. Finally, the simulation results show that the WLS strategy is very effective to standardize multimodal data, and the WLS-SVDD monitoring method has great advantages over the traditional SVDD and PCA combined with a local standardization strategy(LNS-PCA) in multi-mode process monitoring.
文摘Based on the analysis of B3G evolution, the base-band processing chips for mobile terminals are introduced. Key technologies for multi-mode mobile terminal base-band chips are discussed. Terminal technologies are thought to be the key of B3G, and terminal base-band chips are regarded as the core of terminal technologies. Therefore, a unified wireless development platform is required for the R&D of multi-mode mobile terminal base-band processing chips.
基金supported by the Science and Technology Project of State Grid Corporation of China under Grant Number 52094021N010 (5400202199534A-0-5-ZN)
文摘Multi-mode power internet of things(PIoT)combines various communication media to provide spatio-temporal coverage for low-carbon operation in smart park.Edge-end collaboration is feasible to achieve the full utilization of heterogeneous resources and anti-eavesdropping.However,edge-end collaboration-based multi-mode PIoT faces challenges of mutual contradiction in communication and security quality of service(QoS)guarantee,inadaptability of resource management,and multi-mode access conflict.We propose an Adaptive learning based delAysensitive and seCure Edge-End Collaboration algorithm(ACE_(2))to optimize multi-mode channel selection and split device power into artificial noise(AN)transmission and data transmission for secure data delivery.ACE_(2) can achieve multi-attribute QoS guarantee,adaptive resource management and security enhancement,and access conflict elimination with the combined power of deep actor-critic(DAC),“win or learn fast(WoLF)”mechanism,and edge-end collaboration.Simulations demonstrate its superior performance in queuing delay,energy consumption,secrecy capacity,and adaptability to differentiated low-carbon services.
文摘This paper studies the application of mathematical models to analyze the vortex-induced vibrations of the tendons of a given TLP along the Indian coastline, by using an analytical approach, using MATLAB. The tendon is subjected to a steady current load, which causes vortex-shedding downstream, leading to cross-flow vibrations. The magnitude of the excitation(lift and drag coefficients) depends on the vortex-shedding frequency. The resulting vibration is studied for possible resonant behavior. The excitation force is quantified empirically, the added mass by potential flow hydrodynamics, and the vibration by normal mode summation method. Non-linear viscous damping of the water is considered. The non-linear oscillations are studied by the phase-plane method, investigating the limit-cycle oscillations. The stable/unstable regions of the dynamic behavior are demarcated. The modal contribution to the total deflection is studied to establish the possibility of resonance of one of the wet modes with the vortex-shedding frequency.
基金This research was jointly supported by the National Basic Research Program of China under contract N0.2005CB422308the National Natural Science Foundation of China under Contract N0.40545018the National Key laboratory of Remote Sensing Sciences.
文摘The variability of the air-sea system in the low-frequency time domain can be decomposed into several systematic climate modes, namely, the decadal variability (DV) mode, the El Nino Southem Oscillation (ENSO) mode, the annual cycle (AC) mode, the semiannual cycle ( SC ) mode and the intraseasonal variability ( ISV ) mode. The combination of these primary modes in the air - sea system orchestrates a complex climate system. The multi-mode low-frequency variability in SST is investigated based on 22 a SST records from 1982 through 2003. The variation of SST in the past two decades undergoes a different combination of these dominant climate modes over different regions, which leads to an interesting new classification of the global ocean based on the relative importance of these modes. The new classification can provide ideal locations for better monitoring of these low-frequency modes in the scientific proof sense. Moreover, two no-annual variation and 14 no-semiannual variation oceanic points, termed annual and semiannual amphidromes, have been well defined in the AC and SC phase maps. The formation of these nodal points is attributed to the couplings of climate modes in EOF analysis results.
基金This work was supported by the National Key Research and Development Program of China(Grant No.2016YFF0203002)the National Natural Science Foundation of China(Grant No.11702051)+1 种基金China Post-doctoral Science Foundation(Grant No.2017M610176)the Fundamental Research Funds for the Central Universities(DUT16ZD214).
文摘Probability-based diagnostic imaging(PDI)is one of the most well-known damage identification methods using guided waves.It is usually applied to diagnose damage in plates.The previous studies were dependent on the certain damage index(DI)which is always calculated from the guided wave signals.In conventional methods,DI is simply defined by comparing the real-time data with the baseline data as reference.However,the baseline signal is easily affected by varying environmental conditions of structures.In this paper,a reference-free diagnostic imaging method is developed to avoid the influence of environmental factors,such as temperature and load conditions.The DI is defined based on the mode conversion of multi-mode guided waves with realtime signals without baseline signals.To improve the accuracy of diagnosis,two terms are included in the reference-free DI.One is called energy DI,which is defined based on the feature of signal energy.The other is called correlation DI and is defined based on the correlation coefficient.Then the PDI algorithm can be carried out instantaneously according to the reference-free DI.The real-time signals which are used to calculate DI are collected by the piezoelectric lead zirconate titanate(PZT)transducers placed on both sides of a plate.The numerical simulations by the finite element(FE)method on aluminum plates with PZT arrays are performed to validate the effectiveness of the reference-free damage diagnostic imaging.The approach is validated by two different arrays:a circle network and a square network.The results of diagnostic imaging are demonstrated and discussed in this paper.Furthermore,the advantage of reference-free DI is investigated by comparing the accuracy of defined reference-free DI and energy DI.
基金Supported by the National Nature Science Foundation of China(No.61674037)the Priority Academic Program Development of Jiangsu Higher Education Institutions,the National Power Grid Corp Science and Technology Project(No.SGTYHT/16-JS-198)the State Grid Nanjing Power Supply Company Project(No.1701052)
文摘This paper presents a reconfigurable RF front-end for multi-mode multi-standard(MMMS) applications. The designed RF front-end is fabricated in 0.18 μm RF CMOS technology. The low noise characteristic is achieved by the noise canceling technique while the bandwidth is enhanced by gate inductive peaking technique. Measurement results show that, while the input frequency ranges from 100 MHz to 2.9 GHz, the proposed reconfigurable RF front-end achieves a controllable voltage conversion gain(VCG) from 18 dB to 39 dB. The measured maximum input third intercept point(IIP3) is-4.9 dBm and the minimum noise figure(NF) is 4.6 dB. The consumed current ranges from 16 mA to 26.5 mA from a 1.8 V supply voltage. The chip occupies an area of 1.17 mm^2 including pads.
基金supported by the National Natural Science Foundation of China(71171038)
文摘A memetic algorithm (MA) for a multi-mode resourceconstrained project scheduling problem (MRCPSP) is proposed. We use a new fitness function and two very effective local search procedures in the proposed MA. The fitness function makes use of a mechanism called "strategic oscillation" to make the search process have a higher probability to visit solutions around a "feasible boundary". One of the local search procedures aims at improving the lower bound of project makespan to be less than a known upper bound, and another aims at improving a solution of an MRCPSP instance accepting infeasible solutions based on the new fitness function in the search process. A detailed computational experiment is set up using instances from the problem instance library PSPLIB. Computational results show that the proposed MA is very competitive with the state-of-the-art algorithms. The MA obtains improved solutions for one instance of set J30.
基金This work was supported in part by the Ministry National Key Research and Development Project(Grant No.2020AAA0108101)the National Natural Science Foundation of China(Grants No.62125101,62341101,62001018,and 62301011)+1 种基金Shandong Natural Science Foundation(Grant No.ZR2023YQ058)the New Cornerstone Science Foundation through the XPLORER PRIZE.The authors would like to thank Mengyuan Lu and Zengrui Han for their help in the construction of electromagnetic space in Wireless InSite simulation platform and Weibo Wen,Qi Duan,and Yong Yu for their help in the construction of phys ical space in AirSim simulation platform.
文摘The sixth generation(6G)of mobile communication system is witnessing a new paradigm shift,i.e.,integrated sensing-communication system.A comprehensive dataset is a prerequisite for 6G integrated sensing-communication research.This paper develops a novel simulation dataset,named M3SC,for mixed multi-modal(MMM)sensing-communication integration,and the generation framework of the M3SC dataset is further given.To obtain multimodal sensory data in physical space and communication data in electromagnetic space,we utilize Air-Sim and WaveFarer to collect multi-modal sensory data and exploit Wireless InSite to collect communication data.Furthermore,the in-depth integration and precise alignment of AirSim,WaveFarer,andWireless InSite are achieved.The M3SC dataset covers various weather conditions,multiplex frequency bands,and different times of the day.Currently,the M3SC dataset contains 1500 snapshots,including 80 RGB images,160 depth maps,80 LiDAR point clouds,256 sets of mmWave waveforms with 8 radar point clouds,and 72 channel impulse response(CIR)matrices per snapshot,thus totaling 120,000 RGB images,240,000 depth maps,120,000 LiDAR point clouds,384,000 sets of mmWave waveforms with 12,000 radar point clouds,and 108,000 CIR matrices.The data processing result presents the multi-modal sensory information and communication channel statistical properties.Finally,the MMM sensing-communication application,which can be supported by the M3SC dataset,is discussed.
文摘Most visual privacy protection methods only hide the identity information of the face images,but the expression,behavior and some other information,which are of great significant in the live broadcast and other scenarios,are also destroyed by the privacy protection process.To this end,this paper introduces a method to remove the identity information while preserving the expression information by performing multi-mode discriminant analysis on the images normalized with AAM algorithm.The face images are decomposed into mutually orthogonal subspaces corresponding to face attributes such as gender,race and expression,each of which owns related characteristic parameters.Then,the expression parameter is preserves to keep the facial expression information while others parameters,including gender and race,are modified to protect face privacy.The experiments show that this method yields well performance on both data utility and privacy protection.