Dempster-Shafer evidence theory is broadly employed in the research of multi-source information fusion.Nevertheless,when fusing highly conflicting evidence it may pro-duce counterintuitive outcomes.To address this iss...Dempster-Shafer evidence theory is broadly employed in the research of multi-source information fusion.Nevertheless,when fusing highly conflicting evidence it may pro-duce counterintuitive outcomes.To address this issue,a fusion approach based on a newly defined belief exponential diver-gence and Deng entropy is proposed.First,a belief exponential divergence is proposed as the conflict measurement between evidences.Then,the credibility of each evidence is calculated.Afterwards,the Deng entropy is used to calculate information volume to determine the uncertainty of evidence.Then,the weight of evidence is calculated by integrating the credibility and uncertainty of each evidence.Ultimately,initial evidences are amended and fused using Dempster’s rule of combination.The effectiveness of this approach in addressing the fusion of three typical conflict paradoxes is demonstrated by arithmetic exam-ples.Additionally,the proposed approach is applied to aerial tar-get recognition and iris dataset-based classification to validate its efficacy.Results indicate that the proposed approach can enhance the accuracy of target recognition and effectively address the issue of fusing conflicting evidences.展开更多
For some important object recognition applications such as intelligent robots and unmanned driving, images are collected on a consecutive basis and associated among themselves, besides, the scenes have steady prior fe...For some important object recognition applications such as intelligent robots and unmanned driving, images are collected on a consecutive basis and associated among themselves, besides, the scenes have steady prior features. Yet existing technologies do not take full advantage of this information. In order to take object recognition further than existing algorithms in the above application, an object recognition method that fuses temporal sequence with scene priori information is proposed. This method first employs YOLOv3 as the basic algorithm to recognize objects in single-frame images, then the DeepSort algorithm to establish association among potential objects recognized in images of different moments, and finally the confidence fusion method and temporal boundary processing method designed herein to fuse, at the decision level, temporal sequence information with scene priori information. Experiments using public datasets and self-built industrial scene datasets show that due to the expansion of information sources, the quality of single-frame images has less impact on the recognition results, whereby the object recognition is greatly improved. It is presented herein as a widely applicable framework for the fusion of information under multiple classes. All the object recognition algorithms that output object class, location information and recognition confidence at the same time can be integrated into this information fusion framework to improve performance.展开更多
For milling tool life prediction and health management,accurate extraction and dimensionality reduction of its tool wear features are the key to reduce prediction errors.In this paper,we adopt multi-source information...For milling tool life prediction and health management,accurate extraction and dimensionality reduction of its tool wear features are the key to reduce prediction errors.In this paper,we adopt multi-source information fusion technology to extract and fuse the features of cutting vibration signal,cutting force signal and acoustic emission signal in time domain,frequency domain and time-frequency domain,and downscale the sample features by Pearson correlation coefficient to construct a sample data set;then we propose a tool life prediction model based on CNN-SVM optimized by genetic algorithm(GA),which uses CNN convolutional neural network as the feature learner and SVM support vector machine as the trainer for regression prediction.The results show that the improved model in this paper can effectively predict the tool life with better generalization ability,faster network fitting,and 99.85%prediction accuracy.And compared with the BP model,CNN model,SVM model and CNN-SVM model,the performance of the coefficient of determination R2 metric improved by 4.88%,2.96%,2.53%and 1.34%,respectively.展开更多
For a single-structure deep learning fault diagnosis model,its disadvantages are an insufficient feature extraction and weak fault classification capability.This paper proposes a multi-scale deep feature fusion intell...For a single-structure deep learning fault diagnosis model,its disadvantages are an insufficient feature extraction and weak fault classification capability.This paper proposes a multi-scale deep feature fusion intelligent fault diagnosis method based on information entropy.First,a normal autoencoder,denoising autoencoder,sparse autoencoder,and contractive autoencoder are used in parallel to construct a multi-scale deep neural network feature extraction structure.A deep feature fusion strategy based on information entropy is proposed to obtain low-dimensional features and ensure the robustness of the model and the quality of deep features.Finally,the advantage of the deep belief network probability model is used as the fault classifier to identify the faults.The effectiveness of the proposed method was verified by a gearbox test-bed.Experimental results show that,compared with traditional and existing intelligent fault diagnosis methods,the proposed method can obtain representative information and features from the raw data with higher classification accuracy.展开更多
For reservoirs with complex non-Gaussian geological characteristics,such as carbonate reservoirs or reservoirs with sedimentary facies distribution,it is difficult to implement history matching directly,especially for...For reservoirs with complex non-Gaussian geological characteristics,such as carbonate reservoirs or reservoirs with sedimentary facies distribution,it is difficult to implement history matching directly,especially for the ensemble-based data assimilation methods.In this paper,we propose a multi-source information fused generative adversarial network(MSIGAN)model,which is used for parameterization of the complex geologies.In MSIGAN,various information such as facies distribution,microseismic,and inter-well connectivity,can be integrated to learn the geological features.And two major generative models in deep learning,variational autoencoder(VAE)and generative adversarial network(GAN)are combined in our model.Then the proposed MSIGAN model is integrated into the ensemble smoother with multiple data assimilation(ESMDA)method to conduct history matching.We tested the proposed method on two reservoir models with fluvial facies.The experimental results show that the proposed MSIGAN model can effectively learn the complex geological features,which can promote the accuracy of history matching.展开更多
The learning status of learners directly affects the quality of learning.Compared with offline teachers,it is difficult for online teachers to capture the learning status of students in the whole class,and it is even ...The learning status of learners directly affects the quality of learning.Compared with offline teachers,it is difficult for online teachers to capture the learning status of students in the whole class,and it is even more difficult to continue to pay attention to studentswhile teaching.Therefore,this paper proposes an online learning state analysis model based on a convolutional neural network and multi-dimensional information fusion.Specifically,a facial expression recognition model and an eye state recognition model are constructed to detect students’emotions and fatigue,respectively.By integrating the detected data with the homework test score data after online learning,an analysis model of students’online learning status is constructed.According to the PAD model,the learning state is expressed as three dimensions of students’understanding,engagement and interest,and then analyzed from multiple perspectives.Finally,the proposed model is applied to actual teaching,and procedural analysis of 5 different types of online classroom learners is carried out,and the validity of the model is verified by comparing with the results of the manual analysis.展开更多
Multi-source information fusion (MSIF) is imported into structural damage diagnosis methods to improve the validity of damage detection. After the introduction of the basic theory, the function model, classification...Multi-source information fusion (MSIF) is imported into structural damage diagnosis methods to improve the validity of damage detection. After the introduction of the basic theory, the function model, classifications and mathematical methods of MSIF, a structural damage detection method based on MSIF is presented, which is to fuse two or more damage character vectors from different structural damage diagnosis methods on the character-level. In an experiment of concrete plates, modal information is measured and analyzed. The structural damage detection method based on MSIF is taken to localize cracks of concrete plates and it is proved to be effective. Results of damage detection by the method based on MSIF are compared with those from the modal strain energy method and the flexibility method. Damage, which can hardly be detected by using the single damage identification method, can be diagnosed by the damage detection method based on the character-level MSIF technique. Meanwhile multi-location damage can be identified by the method based on MSIF. This method is sensitive to structural damage and different mathematical methods for MSIF have different preconditions and applicabilities for diversified structures. How to choose mathematical methods for MSIF should be discussed in detail in health monitoring systems of actual structures.展开更多
With the emergence and development of social networks,people can stay in touch with friends,family,and colleagues more quickly and conveniently,regardless of their location.This ubiquitous digital internet environment...With the emergence and development of social networks,people can stay in touch with friends,family,and colleagues more quickly and conveniently,regardless of their location.This ubiquitous digital internet environment has also led to large-scale disclosure of personal privacy.Due to the complexity and subtlety of sensitive information,traditional sensitive information identification technologies cannot thoroughly address the characteristics of each piece of data,thus weakening the deep connections between text and images.In this context,this paper adopts the CLIP model as a modality discriminator.By using comparative learning between sensitive image descriptions and images,the similarity between the images and the sensitive descriptions is obtained to determine whether the images contain sensitive information.This provides the basis for identifying sensitive information using different modalities.Specifically,if the original data does not contain sensitive information,only single-modality text-sensitive information identification is performed;if the original data contains sensitive information,multimodality sensitive information identification is conducted.This approach allows for differentiated processing of each piece of data,thereby achieving more accurate sensitive information identification.The aforementioned modality discriminator can address the limitations of existing sensitive information identification technologies,making the identification of sensitive information from the original data more appropriate and precise.展开更多
Personality distinguishes individuals’ patterns of feeling, thinking,and behaving. Predicting personality from small video series is an excitingresearch area in computer vision. The majority of the existing research ...Personality distinguishes individuals’ patterns of feeling, thinking,and behaving. Predicting personality from small video series is an excitingresearch area in computer vision. The majority of the existing research concludespreliminary results to get immense knowledge from visual and Audio(sound) modality. To overcome the deficiency, we proposed the Deep BimodalFusion (DBF) approach to predict five traits of personality-agreeableness,extraversion, openness, conscientiousness and neuroticism. In the proposedframework, regarding visual modality, the modified convolution neural networks(CNN), more specifically Descriptor Aggregator Model (DAN) areused to attain significant visual modality. The proposed model extracts audiorepresentations for greater efficiency to construct the long short-termmemory(LSTM) for the audio modality. Moreover, employing modality-based neuralnetworks allows this framework to independently determine the traits beforecombining them with weighted fusion to achieve a conclusive prediction of thegiven traits. The proposed approach attains the optimal mean accuracy score,which is 0.9183. It is achieved based on the average of five personality traitsand is thus better than previously proposed frameworks.展开更多
This paper addresses the challenge of accurately and timely determining the position of a train,with specific consideration given to the integration of the global navigation satellite system(GNSS)and inertial navigati...This paper addresses the challenge of accurately and timely determining the position of a train,with specific consideration given to the integration of the global navigation satellite system(GNSS)and inertial navigation system(INS).To overcome the increasing errors in the INS during interruptions in GNSS signals,as well as the uncertainty associated with process and measurement noise,a deep learning-based method for train positioning is proposed.This method combines convolutional neural networks(CNN),long short-term memory(LSTM),and the invariant extended Kalman filter(IEKF)to enhance the perception of train positions.It effectively handles GNSS signal interruptions and mitigates the impact of noise.Experimental evaluation and comparisons with existing approaches are provided to illustrate the effectiveness and robustness of the proposed method.展开更多
Nowadays short texts can be widely found in various social data in relation to the 5G-enabled Internet of Things (IoT). Short text classification is a challenging task due to its sparsity and the lack of context. Prev...Nowadays short texts can be widely found in various social data in relation to the 5G-enabled Internet of Things (IoT). Short text classification is a challenging task due to its sparsity and the lack of context. Previous studies mainly tackle these problems by enhancing the semantic information or the statistical information individually. However, the improvement achieved by a single type of information is limited, while fusing various information may help to improve the classification accuracy more effectively. To fuse various information for short text classification, this article proposes a feature fusion method that integrates the statistical feature and the comprehensive semantic feature together by using the weighting mechanism and deep learning models. In the proposed method, we apply Bidirectional Encoder Representations from Transformers (BERT) to generate word vectors on the sentence level automatically, and then obtain the statistical feature, the local semantic feature and the overall semantic feature using Term Frequency-Inverse Document Frequency (TF-IDF) weighting approach, Convolutional Neural Network (CNN) and Bidirectional Gate Recurrent Unit (BiGRU). Then, the fusion feature is accordingly obtained for classification. Experiments are conducted on five popular short text classification datasets and a 5G-enabled IoT social dataset and the results show that our proposed method effectively improves the classification performance.展开更多
The recent COVID-19 pandemic caused by the novel coronavirus,severe acute respiratory syndrome coronavirus 2(SARS-CoV-2),has had a significant impact on human life and the economy around the world.A reverse transcript...The recent COVID-19 pandemic caused by the novel coronavirus,severe acute respiratory syndrome coronavirus 2(SARS-CoV-2),has had a significant impact on human life and the economy around the world.A reverse transcription polymerase chain reaction(RT-PCR)test is used to screen for this disease,but its low sensitivity means that it is not sufficient for early detection and treatment.As RT-PCR is a time-consuming procedure,there is interest in the introduction of automated techniques for diagnosis.Deep learning has a key role to play in the field of medical imaging.The most important issue in this area is the choice of key features.Here,we propose a set of deep learning features based on a system for automated classification of computed tomography(CT)images to identify COVID-19.Initially,this method was used to prepare a database of three classes:Pneumonia,COVID19,and Healthy.The dataset consisted of 6000 CT images refined by a hybrid contrast stretching approach.In the next step,two advanced deep learning models(ResNet50 and DarkNet53)were fine-tuned and trained through transfer learning.The features were extracted from the second last feature layer of both models and further optimized using a hybrid optimization approach.For each deep model,the Rao-1 algorithm and the PSO algorithm were combined in the hybrid approach.Later,the selected features were merged using the new minimum parallel distance non-redundant(PMDNR)approach.The final fused vector was finally classified using the extreme machine classifier.The experimental process was carried out on a set of prepared data with an overall accuracy of 95.6%.Comparing the different classification algorithms at the different levels of the features demonstrated the reliability of the proposed framework.展开更多
Aerial scene recognition(ASR)has attracted great attention due to its increasingly essential applications.Most of the ASR methods adopt the multi‐scale architecture because both global and local features play great r...Aerial scene recognition(ASR)has attracted great attention due to its increasingly essential applications.Most of the ASR methods adopt the multi‐scale architecture because both global and local features play great roles in ASR.However,the existing multi‐scale methods neglect the effective interactions among different scales and various spatial locations when fusing global and local features,leading to a limited ability to deal with challenges of large‐scale variation and complex background in aerial scene images.In addition,existing methods may suffer from poor generalisations due to millions of to‐belearnt parameters and inconsistent predictions between global and local features.To tackle these problems,this study proposes a scale‐wise interaction fusion and knowledge distillation(SIF‐KD)network for learning robust and discriminative features with scaleinvariance and background‐independent information.The main highlights of this study include two aspects.On the one hand,a global‐local features collaborative learning scheme is devised for extracting scale‐invariance features so as to tackle the large‐scale variation problem in aerial scene images.Specifically,a plug‐and‐play multi‐scale context attention fusion module is proposed for collaboratively fusing the context information between global and local features.On the other hand,a scale‐wise knowledge distillation scheme is proposed to produce more consistent predictions by distilling the predictive distribution between different scales during training.Comprehensive experimental results show the proposed SIF‐KD network achieves the best overall accuracy with 99.68%,98.74%and 95.47%on the UCM,AID and NWPU‐RESISC45 datasets,respectively,compared with state of the arts.展开更多
Diagnosing gastrointestinal cancer by classical means is a hazardous procedure.Years have witnessed several computerized solutions for stomach disease detection and classification.However,the existing techniques faced...Diagnosing gastrointestinal cancer by classical means is a hazardous procedure.Years have witnessed several computerized solutions for stomach disease detection and classification.However,the existing techniques faced challenges,such as irrelevant feature extraction,high similarity among different disease symptoms,and the least-important features from a single source.This paper designed a new deep learning-based architecture based on the fusion of two models,Residual blocks and Auto Encoder.First,the Hyper-Kvasir dataset was employed to evaluate the proposed work.The research selected a pre-trained convolutional neural network(CNN)model and improved it with several residual blocks.This process aims to improve the learning capability of deep models and lessen the number of parameters.Besides,this article designed an Auto-Encoder-based network consisting of five convolutional layers in the encoder stage and five in the decoder phase.The research selected the global average pooling and convolutional layers for the feature extraction optimized by a hybrid Marine Predator optimization and Slime Mould optimization algorithm.These features of both models are fused using a novel fusion technique that is later classified using the Artificial Neural Network classifier.The experiment worked on the HyperKvasir dataset,which consists of 23 stomach-infected classes.At last,the proposed method obtained an improved accuracy of 93.90%on this dataset.Comparison is also conducted with some recent techniques and shows that the proposed method’s accuracy is improved.展开更多
Image fusion refers to extracting meaningful information from images of different sources or modalities,and then fusing them to generate more informative images that are beneficial for subsequent applications.In recen...Image fusion refers to extracting meaningful information from images of different sources or modalities,and then fusing them to generate more informative images that are beneficial for subsequent applications.In recent years,the growing data and computing resources have promoted the development of deep learning,and image fusion technology has continued to spawn new deep learning fusion methods based on traditional fusion methods.However,high-speed railroads,as an important part of life,have their unique industry characteristics of image data,which leads to different image fusion techniques with different fusion effects in high-speed railway scenes.This research work first introduces the mainstream technology classification of image fusion,further describes the downstream tasks that image fusion techniques may combine within high-speed railway scenes,and introduces the evaluation metrics of image fusion,followed by a series of subjective and objective experiments to completely evaluate the performance level of each image fusion method in different traffic scenes,and finally provides some possible future image fusion in the field of rail transportation of research.展开更多
基金supported by the National Natural Science Foundation of China(61903305,62073267)the Fundamental Research Funds for the Central Universities(HXGJXM202214).
文摘Dempster-Shafer evidence theory is broadly employed in the research of multi-source information fusion.Nevertheless,when fusing highly conflicting evidence it may pro-duce counterintuitive outcomes.To address this issue,a fusion approach based on a newly defined belief exponential diver-gence and Deng entropy is proposed.First,a belief exponential divergence is proposed as the conflict measurement between evidences.Then,the credibility of each evidence is calculated.Afterwards,the Deng entropy is used to calculate information volume to determine the uncertainty of evidence.Then,the weight of evidence is calculated by integrating the credibility and uncertainty of each evidence.Ultimately,initial evidences are amended and fused using Dempster’s rule of combination.The effectiveness of this approach in addressing the fusion of three typical conflict paradoxes is demonstrated by arithmetic exam-ples.Additionally,the proposed approach is applied to aerial tar-get recognition and iris dataset-based classification to validate its efficacy.Results indicate that the proposed approach can enhance the accuracy of target recognition and effectively address the issue of fusing conflicting evidences.
文摘For some important object recognition applications such as intelligent robots and unmanned driving, images are collected on a consecutive basis and associated among themselves, besides, the scenes have steady prior features. Yet existing technologies do not take full advantage of this information. In order to take object recognition further than existing algorithms in the above application, an object recognition method that fuses temporal sequence with scene priori information is proposed. This method first employs YOLOv3 as the basic algorithm to recognize objects in single-frame images, then the DeepSort algorithm to establish association among potential objects recognized in images of different moments, and finally the confidence fusion method and temporal boundary processing method designed herein to fuse, at the decision level, temporal sequence information with scene priori information. Experiments using public datasets and self-built industrial scene datasets show that due to the expansion of information sources, the quality of single-frame images has less impact on the recognition results, whereby the object recognition is greatly improved. It is presented herein as a widely applicable framework for the fusion of information under multiple classes. All the object recognition algorithms that output object class, location information and recognition confidence at the same time can be integrated into this information fusion framework to improve performance.
基金financed with the means of Basic Scientific Research Youth Program of Education Department of Liaoning Province,No.LJKQZ2021185Yingkou Enterprise and Doctor Innovation Program (QB-2021-05).
文摘For milling tool life prediction and health management,accurate extraction and dimensionality reduction of its tool wear features are the key to reduce prediction errors.In this paper,we adopt multi-source information fusion technology to extract and fuse the features of cutting vibration signal,cutting force signal and acoustic emission signal in time domain,frequency domain and time-frequency domain,and downscale the sample features by Pearson correlation coefficient to construct a sample data set;then we propose a tool life prediction model based on CNN-SVM optimized by genetic algorithm(GA),which uses CNN convolutional neural network as the feature learner and SVM support vector machine as the trainer for regression prediction.The results show that the improved model in this paper can effectively predict the tool life with better generalization ability,faster network fitting,and 99.85%prediction accuracy.And compared with the BP model,CNN model,SVM model and CNN-SVM model,the performance of the coefficient of determination R2 metric improved by 4.88%,2.96%,2.53%and 1.34%,respectively.
基金Supported by National Natural Science Foundation of China and Civil Aviation Administration of China Joint Funded Project(Grant No.U1733108)Key Project of Tianjin Science and Technology Support Program(Grant No.16YFZCSY00860).
文摘For a single-structure deep learning fault diagnosis model,its disadvantages are an insufficient feature extraction and weak fault classification capability.This paper proposes a multi-scale deep feature fusion intelligent fault diagnosis method based on information entropy.First,a normal autoencoder,denoising autoencoder,sparse autoencoder,and contractive autoencoder are used in parallel to construct a multi-scale deep neural network feature extraction structure.A deep feature fusion strategy based on information entropy is proposed to obtain low-dimensional features and ensure the robustness of the model and the quality of deep features.Finally,the advantage of the deep belief network probability model is used as the fault classifier to identify the faults.The effectiveness of the proposed method was verified by a gearbox test-bed.Experimental results show that,compared with traditional and existing intelligent fault diagnosis methods,the proposed method can obtain representative information and features from the raw data with higher classification accuracy.
基金supported by the National Natural Science Foundation of China under Grant 51722406,52074340,and 51874335the Shandong Provincial Natural Science Foundation under Grant JQ201808+5 种基金The Fundamental Research Funds for the Central Universities under Grant 18CX02097Athe Major Scientific and Technological Projects of CNPC under Grant ZD2019-183-008the Science and Technology Support Plan for Youth Innovation of University in Shandong Province under Grant 2019KJH002the National Research Council of Science and Technology Major Project of China under Grant 2016ZX05025001-006111 Project under Grant B08028Sinopec Science and Technology Project under Grant P20050-1
文摘For reservoirs with complex non-Gaussian geological characteristics,such as carbonate reservoirs or reservoirs with sedimentary facies distribution,it is difficult to implement history matching directly,especially for the ensemble-based data assimilation methods.In this paper,we propose a multi-source information fused generative adversarial network(MSIGAN)model,which is used for parameterization of the complex geologies.In MSIGAN,various information such as facies distribution,microseismic,and inter-well connectivity,can be integrated to learn the geological features.And two major generative models in deep learning,variational autoencoder(VAE)and generative adversarial network(GAN)are combined in our model.Then the proposed MSIGAN model is integrated into the ensemble smoother with multiple data assimilation(ESMDA)method to conduct history matching.We tested the proposed method on two reservoir models with fluvial facies.The experimental results show that the proposed MSIGAN model can effectively learn the complex geological features,which can promote the accuracy of history matching.
基金supported by the Chongqing Normal University Graduate Scientific Research Innovation Project (Grants YZH21014 and YZH21010).
文摘The learning status of learners directly affects the quality of learning.Compared with offline teachers,it is difficult for online teachers to capture the learning status of students in the whole class,and it is even more difficult to continue to pay attention to studentswhile teaching.Therefore,this paper proposes an online learning state analysis model based on a convolutional neural network and multi-dimensional information fusion.Specifically,a facial expression recognition model and an eye state recognition model are constructed to detect students’emotions and fatigue,respectively.By integrating the detected data with the homework test score data after online learning,an analysis model of students’online learning status is constructed.According to the PAD model,the learning state is expressed as three dimensions of students’understanding,engagement and interest,and then analyzed from multiple perspectives.Finally,the proposed model is applied to actual teaching,and procedural analysis of 5 different types of online classroom learners is carried out,and the validity of the model is verified by comparing with the results of the manual analysis.
基金The National High Technology Research and Develop-ment Program of China(863Program)(No.2006AA04Z416)the Na-tional Science Fund for Distinguished Young Scholars(No.50725828)the Excellent Dissertation Program for Doctoral Degree of Southeast University(No.0705)
文摘Multi-source information fusion (MSIF) is imported into structural damage diagnosis methods to improve the validity of damage detection. After the introduction of the basic theory, the function model, classifications and mathematical methods of MSIF, a structural damage detection method based on MSIF is presented, which is to fuse two or more damage character vectors from different structural damage diagnosis methods on the character-level. In an experiment of concrete plates, modal information is measured and analyzed. The structural damage detection method based on MSIF is taken to localize cracks of concrete plates and it is proved to be effective. Results of damage detection by the method based on MSIF are compared with those from the modal strain energy method and the flexibility method. Damage, which can hardly be detected by using the single damage identification method, can be diagnosed by the damage detection method based on the character-level MSIF technique. Meanwhile multi-location damage can be identified by the method based on MSIF. This method is sensitive to structural damage and different mathematical methods for MSIF have different preconditions and applicabilities for diversified structures. How to choose mathematical methods for MSIF should be discussed in detail in health monitoring systems of actual structures.
基金supported by the National Natural Science Foundation of China(No.62302540),with author Fangfang Shan for more information,please visit their website at https://www.nsfc.gov.cn/(accessed on 05 June 2024)Additionally,it is also funded by the Open Foundation of Henan Key Laboratory of Cyberspace Situation Awareness(No.HNTS2022020),where Fangfang Shan is an author.Further details can be found at http://xt.hnkjt.gov.cn/data/pingtai/(accessed on 05 June 2024)the Natural Science Foundation of Henan Province Youth Science Fund Project(No.232300420422),and for more information,you can visit https://kjt.henan.gov.cn(accessed on 05 June 2024).
文摘With the emergence and development of social networks,people can stay in touch with friends,family,and colleagues more quickly and conveniently,regardless of their location.This ubiquitous digital internet environment has also led to large-scale disclosure of personal privacy.Due to the complexity and subtlety of sensitive information,traditional sensitive information identification technologies cannot thoroughly address the characteristics of each piece of data,thus weakening the deep connections between text and images.In this context,this paper adopts the CLIP model as a modality discriminator.By using comparative learning between sensitive image descriptions and images,the similarity between the images and the sensitive descriptions is obtained to determine whether the images contain sensitive information.This provides the basis for identifying sensitive information using different modalities.Specifically,if the original data does not contain sensitive information,only single-modality text-sensitive information identification is performed;if the original data contains sensitive information,multimodality sensitive information identification is conducted.This approach allows for differentiated processing of each piece of data,thereby achieving more accurate sensitive information identification.The aforementioned modality discriminator can address the limitations of existing sensitive information identification technologies,making the identification of sensitive information from the original data more appropriate and precise.
文摘Personality distinguishes individuals’ patterns of feeling, thinking,and behaving. Predicting personality from small video series is an excitingresearch area in computer vision. The majority of the existing research concludespreliminary results to get immense knowledge from visual and Audio(sound) modality. To overcome the deficiency, we proposed the Deep BimodalFusion (DBF) approach to predict five traits of personality-agreeableness,extraversion, openness, conscientiousness and neuroticism. In the proposedframework, regarding visual modality, the modified convolution neural networks(CNN), more specifically Descriptor Aggregator Model (DAN) areused to attain significant visual modality. The proposed model extracts audiorepresentations for greater efficiency to construct the long short-termmemory(LSTM) for the audio modality. Moreover, employing modality-based neuralnetworks allows this framework to independently determine the traits beforecombining them with weighted fusion to achieve a conclusive prediction of thegiven traits. The proposed approach attains the optimal mean accuracy score,which is 0.9183. It is achieved based on the average of five personality traitsand is thus better than previously proposed frameworks.
基金supported by the National Natural Science Foundation of China(Nos.61925302,62273027)the Beijing Natural Science Foundation(L211021).
文摘This paper addresses the challenge of accurately and timely determining the position of a train,with specific consideration given to the integration of the global navigation satellite system(GNSS)and inertial navigation system(INS).To overcome the increasing errors in the INS during interruptions in GNSS signals,as well as the uncertainty associated with process and measurement noise,a deep learning-based method for train positioning is proposed.This method combines convolutional neural networks(CNN),long short-term memory(LSTM),and the invariant extended Kalman filter(IEKF)to enhance the perception of train positions.It effectively handles GNSS signal interruptions and mitigates the impact of noise.Experimental evaluation and comparisons with existing approaches are provided to illustrate the effectiveness and robustness of the proposed method.
基金supported in part by the Beijing Natural Science Foundation under grants M21032 and 19L2029in part by the National Natural Science Foundation of China under grants U1836106 and 81961138010in part by the Scientific and Technological Innovation Foundation of Foshan under grants BK21BF001 and BK20BF010.
文摘Nowadays short texts can be widely found in various social data in relation to the 5G-enabled Internet of Things (IoT). Short text classification is a challenging task due to its sparsity and the lack of context. Previous studies mainly tackle these problems by enhancing the semantic information or the statistical information individually. However, the improvement achieved by a single type of information is limited, while fusing various information may help to improve the classification accuracy more effectively. To fuse various information for short text classification, this article proposes a feature fusion method that integrates the statistical feature and the comprehensive semantic feature together by using the weighting mechanism and deep learning models. In the proposed method, we apply Bidirectional Encoder Representations from Transformers (BERT) to generate word vectors on the sentence level automatically, and then obtain the statistical feature, the local semantic feature and the overall semantic feature using Term Frequency-Inverse Document Frequency (TF-IDF) weighting approach, Convolutional Neural Network (CNN) and Bidirectional Gate Recurrent Unit (BiGRU). Then, the fusion feature is accordingly obtained for classification. Experiments are conducted on five popular short text classification datasets and a 5G-enabled IoT social dataset and the results show that our proposed method effectively improves the classification performance.
基金This research was supported by X-mind Corps program of National Research Foundation of Korea(NRF)funded by the Ministry of Science,ICT(No.2019H1D8A1105622)the Soonchunhyang University Research Fund.
文摘The recent COVID-19 pandemic caused by the novel coronavirus,severe acute respiratory syndrome coronavirus 2(SARS-CoV-2),has had a significant impact on human life and the economy around the world.A reverse transcription polymerase chain reaction(RT-PCR)test is used to screen for this disease,but its low sensitivity means that it is not sufficient for early detection and treatment.As RT-PCR is a time-consuming procedure,there is interest in the introduction of automated techniques for diagnosis.Deep learning has a key role to play in the field of medical imaging.The most important issue in this area is the choice of key features.Here,we propose a set of deep learning features based on a system for automated classification of computed tomography(CT)images to identify COVID-19.Initially,this method was used to prepare a database of three classes:Pneumonia,COVID19,and Healthy.The dataset consisted of 6000 CT images refined by a hybrid contrast stretching approach.In the next step,two advanced deep learning models(ResNet50 and DarkNet53)were fine-tuned and trained through transfer learning.The features were extracted from the second last feature layer of both models and further optimized using a hybrid optimization approach.For each deep model,the Rao-1 algorithm and the PSO algorithm were combined in the hybrid approach.Later,the selected features were merged using the new minimum parallel distance non-redundant(PMDNR)approach.The final fused vector was finally classified using the extreme machine classifier.The experimental process was carried out on a set of prepared data with an overall accuracy of 95.6%.Comparing the different classification algorithms at the different levels of the features demonstrated the reliability of the proposed framework.
基金supported in part by the National Natural Science Foundation of China under Grant 62201452,2271296 and 62201453in part by the Natural Science Basic Research Programme of Shaanxi under Grant 2022JQ‐592+1 种基金in part by the Special Construction Fund for Key Disciplines of Shaanxi Provincial Higher Education,in part by the Natural Science Basic Research Program of Shaanxi under Grant 2021JC‐47in part by Scientific Research Program Funded by Shaanxi Provincial Education Department under Grant 22JK0568.
文摘Aerial scene recognition(ASR)has attracted great attention due to its increasingly essential applications.Most of the ASR methods adopt the multi‐scale architecture because both global and local features play great roles in ASR.However,the existing multi‐scale methods neglect the effective interactions among different scales and various spatial locations when fusing global and local features,leading to a limited ability to deal with challenges of large‐scale variation and complex background in aerial scene images.In addition,existing methods may suffer from poor generalisations due to millions of to‐belearnt parameters and inconsistent predictions between global and local features.To tackle these problems,this study proposes a scale‐wise interaction fusion and knowledge distillation(SIF‐KD)network for learning robust and discriminative features with scaleinvariance and background‐independent information.The main highlights of this study include two aspects.On the one hand,a global‐local features collaborative learning scheme is devised for extracting scale‐invariance features so as to tackle the large‐scale variation problem in aerial scene images.Specifically,a plug‐and‐play multi‐scale context attention fusion module is proposed for collaboratively fusing the context information between global and local features.On the other hand,a scale‐wise knowledge distillation scheme is proposed to produce more consistent predictions by distilling the predictive distribution between different scales during training.Comprehensive experimental results show the proposed SIF‐KD network achieves the best overall accuracy with 99.68%,98.74%and 95.47%on the UCM,AID and NWPU‐RESISC45 datasets,respectively,compared with state of the arts.
基金supported by“Human Resources Program in Energy Technology”of the Korea Institute of Energy Technology Evaluation and Planning(KETEP),granted financial resources from the Ministry of Trade,Industry&Energy,Republic of Korea(No.20204010600090)Supporting Project Number(PNURSP2023R387),Princess Nourah bint Abdulrahman University,Riyadh,Saudi Arabia.
文摘Diagnosing gastrointestinal cancer by classical means is a hazardous procedure.Years have witnessed several computerized solutions for stomach disease detection and classification.However,the existing techniques faced challenges,such as irrelevant feature extraction,high similarity among different disease symptoms,and the least-important features from a single source.This paper designed a new deep learning-based architecture based on the fusion of two models,Residual blocks and Auto Encoder.First,the Hyper-Kvasir dataset was employed to evaluate the proposed work.The research selected a pre-trained convolutional neural network(CNN)model and improved it with several residual blocks.This process aims to improve the learning capability of deep models and lessen the number of parameters.Besides,this article designed an Auto-Encoder-based network consisting of five convolutional layers in the encoder stage and five in the decoder phase.The research selected the global average pooling and convolutional layers for the feature extraction optimized by a hybrid Marine Predator optimization and Slime Mould optimization algorithm.These features of both models are fused using a novel fusion technique that is later classified using the Artificial Neural Network classifier.The experiment worked on the HyperKvasir dataset,which consists of 23 stomach-infected classes.At last,the proposed method obtained an improved accuracy of 93.90%on this dataset.Comparison is also conducted with some recent techniques and shows that the proposed method’s accuracy is improved.
基金supported in part by the National Key Research and Development Program of China,under Grant 2020YFB2103800.
文摘Image fusion refers to extracting meaningful information from images of different sources or modalities,and then fusing them to generate more informative images that are beneficial for subsequent applications.In recent years,the growing data and computing resources have promoted the development of deep learning,and image fusion technology has continued to spawn new deep learning fusion methods based on traditional fusion methods.However,high-speed railroads,as an important part of life,have their unique industry characteristics of image data,which leads to different image fusion techniques with different fusion effects in high-speed railway scenes.This research work first introduces the mainstream technology classification of image fusion,further describes the downstream tasks that image fusion techniques may combine within high-speed railway scenes,and introduces the evaluation metrics of image fusion,followed by a series of subjective and objective experiments to completely evaluate the performance level of each image fusion method in different traffic scenes,and finally provides some possible future image fusion in the field of rail transportation of research.