期刊文献+
共找到4,249篇文章
< 1 2 213 >
每页显示 20 50 100
Hyperspectral remote sensing identification of marine oil emulsions based on the fusion of spatial and spectral features
1
作者 Xinyue Huang Yi Ma +1 位作者 Zongchen Jiang Junfang Yang 《Acta Oceanologica Sinica》 SCIE CAS CSCD 2024年第3期139-154,共16页
Marine oil spill emulsions are difficult to recover,and the damage to the environment is not easy to eliminate.The use of remote sensing to accurately identify oil spill emulsions is highly important for the protectio... Marine oil spill emulsions are difficult to recover,and the damage to the environment is not easy to eliminate.The use of remote sensing to accurately identify oil spill emulsions is highly important for the protection of marine environments.However,the spectrum of oil emulsions changes due to different water content.Hyperspectral remote sensing and deep learning can use spectral and spatial information to identify different types of oil emulsions.Nonetheless,hyperspectral data can also cause information redundancy,reducing classification accuracy and efficiency,and even overfitting in machine learning models.To address these problems,an oil emulsion deep-learning identification model with spatial-spectral feature fusion is established,and feature bands that can distinguish between crude oil,seawater,water-in-oil emulsion(WO),and oil-in-water emulsion(OW)are filtered based on a standard deviation threshold–mutual information method.Using oil spill airborne hyperspectral data,we conducted identification experiments on oil emulsions in different background waters and under different spatial and temporal conditions,analyzed the transferability of the model,and explored the effects of feature band selection and spectral resolution on the identification of oil emulsions.The results show the following.(1)The standard deviation–mutual information feature selection method is able to effectively extract feature bands that can distinguish between WO,OW,oil slick,and seawater.The number of bands was reduced from 224 to 134 after feature selection on the Airborne Visible Infrared Imaging Spectrometer(AVIRIS)data and from 126 to 100 on the S185 data.(2)With feature selection,the overall accuracy and Kappa of the identification results for the training area are 91.80%and 0.86,respectively,improved by 2.62%and 0.04,and the overall accuracy and Kappa of the identification results for the migration area are 86.53%and 0.80,respectively,improved by 3.45%and 0.05.(3)The oil emulsion identification model has a certain degree of transferability and can effectively identify oil spill emulsions for AVIRIS data at different times and locations,with an overall accuracy of more than 80%,Kappa coefficient of more than 0.7,and F1 score of 0.75 or more for each category.(4)As the spectral resolution decreasing,the model yields different degrees of misclassification for areas with a mixed distribution of oil slick and seawater or mixed distribution of WO and OW.Based on the above experimental results,we demonstrate that the oil emulsion identification model with spatial–spectral feature fusion achieves a high accuracy rate in identifying oil emulsion using airborne hyperspectral data,and can be applied to images under different spatial and temporal conditions.Furthermore,we also elucidate the impact of factors such as spectral resolution and background water bodies on the identification process.These findings provide new reference for future endeavors in automated marine oil spill detection. 展开更多
关键词 oil emulsions IDENTIFICATION hyperspectral remote sensing feature selection convolutional neural network(CNN) spatial-temporal transferability
下载PDF
Olive Leaf Disease Detection via Wavelet Transform and Feature Fusion of Pre-Trained Deep Learning Models
2
作者 Mahmood A.Mahmood Khalaf Alsalem 《Computers, Materials & Continua》 SCIE EI 2024年第3期3431-3448,共18页
Olive trees are susceptible to a variety of diseases that can cause significant crop damage and economic losses.Early detection of these diseases is essential for effective management.We propose a novel transformed wa... Olive trees are susceptible to a variety of diseases that can cause significant crop damage and economic losses.Early detection of these diseases is essential for effective management.We propose a novel transformed wavelet,feature-fused,pre-trained deep learning model for detecting olive leaf diseases.The proposed model combines wavelet transforms with pre-trained deep-learning models to extract discriminative features from olive leaf images.The model has four main phases:preprocessing using data augmentation,three-level wavelet transformation,learning using pre-trained deep learning models,and a fused deep learning model.In the preprocessing phase,the image dataset is augmented using techniques such as resizing,rescaling,flipping,rotation,zooming,and contrasting.In wavelet transformation,the augmented images are decomposed into three frequency levels.Three pre-trained deep learning models,EfficientNet-B7,DenseNet-201,and ResNet-152-V2,are used in the learning phase.The models were trained using the approximate images of the third-level sub-band of the wavelet transform.In the fused phase,the fused model consists of a merge layer,three dense layers,and two dropout layers.The proposed model was evaluated using a dataset of images of healthy and infected olive leaves.It achieved an accuracy of 99.72%in the diagnosis of olive leaf diseases,which exceeds the accuracy of other methods reported in the literature.This finding suggests that our proposed method is a promising tool for the early detection of olive leaf diseases. 展开更多
关键词 Olive leaf diseases wavelet transform deep learning feature fusion
下载PDF
Exploring Sequential Feature Selection in Deep Bi-LSTM Models for Speech Emotion Recognition
3
作者 Fatma Harby Mansor Alohali +1 位作者 Adel Thaljaoui Amira Samy Talaat 《Computers, Materials & Continua》 SCIE EI 2024年第2期2689-2719,共31页
Machine Learning(ML)algorithms play a pivotal role in Speech Emotion Recognition(SER),although they encounter a formidable obstacle in accurately discerning a speaker’s emotional state.The examination of the emotiona... Machine Learning(ML)algorithms play a pivotal role in Speech Emotion Recognition(SER),although they encounter a formidable obstacle in accurately discerning a speaker’s emotional state.The examination of the emotional states of speakers holds significant importance in a range of real-time applications,including but not limited to virtual reality,human-robot interaction,emergency centers,and human behavior assessment.Accurately identifying emotions in the SER process relies on extracting relevant information from audio inputs.Previous studies on SER have predominantly utilized short-time characteristics such as Mel Frequency Cepstral Coefficients(MFCCs)due to their ability to capture the periodic nature of audio signals effectively.Although these traits may improve their ability to perceive and interpret emotional depictions appropriately,MFCCS has some limitations.So this study aims to tackle the aforementioned issue by systematically picking multiple audio cues,enhancing the classifier model’s efficacy in accurately discerning human emotions.The utilized dataset is taken from the EMO-DB database,preprocessing input speech is done using a 2D Convolution Neural Network(CNN)involves applying convolutional operations to spectrograms as they afford a visual representation of the way the audio signal frequency content changes over time.The next step is the spectrogram data normalization which is crucial for Neural Network(NN)training as it aids in faster convergence.Then the five auditory features MFCCs,Chroma,Mel-Spectrogram,Contrast,and Tonnetz are extracted from the spectrogram sequentially.The attitude of feature selection is to retain only dominant features by excluding the irrelevant ones.In this paper,the Sequential Forward Selection(SFS)and Sequential Backward Selection(SBS)techniques were employed for multiple audio cues features selection.Finally,the feature sets composed from the hybrid feature extraction methods are fed into the deep Bidirectional Long Short Term Memory(Bi-LSTM)network to discern emotions.Since the deep Bi-LSTM can hierarchically learn complex features and increases model capacity by achieving more robust temporal modeling,it is more effective than a shallow Bi-LSTM in capturing the intricate tones of emotional content existent in speech signals.The effectiveness and resilience of the proposed SER model were evaluated by experiments,comparing it to state-of-the-art SER techniques.The results indicated that the model achieved accuracy rates of 90.92%,93%,and 92%over the Ryerson Audio-Visual Database of Emotional Speech and Song(RAVDESS),Berlin Database of Emotional Speech(EMO-DB),and The Interactive Emotional Dyadic Motion Capture(IEMOCAP)datasets,respectively.These findings signify a prominent enhancement in the ability to emotional depictions identification in speech,showcasing the potential of the proposed model in advancing the SER field. 展开更多
关键词 Artificial intelligence application multi features sequential selection speech emotion recognition deep Bi-LSTM
下载PDF
Nonparametric Statistical Feature Scaling Based Quadratic Regressive Convolution Deep Neural Network for Software Fault Prediction
4
作者 Sureka Sivavelu Venkatesh Palanisamy 《Computers, Materials & Continua》 SCIE EI 2024年第3期3469-3487,共19页
The development of defect prediction plays a significant role in improving software quality. Such predictions are used to identify defective modules before the testing and to minimize the time and cost. The software w... The development of defect prediction plays a significant role in improving software quality. Such predictions are used to identify defective modules before the testing and to minimize the time and cost. The software with defects negatively impacts operational costs and finally affects customer satisfaction. Numerous approaches exist to predict software defects. However, the timely and accurate software bugs are the major challenging issues. To improve the timely and accurate software defect prediction, a novel technique called Nonparametric Statistical feature scaled QuAdratic regressive convolution Deep nEural Network (SQADEN) is introduced. The proposed SQADEN technique mainly includes two major processes namely metric or feature selection and classification. First, the SQADEN uses the nonparametric statistical Torgerson–Gower scaling technique for identifying the relevant software metrics by measuring the similarity using the dice coefficient. The feature selection process is used to minimize the time complexity of software fault prediction. With the selected metrics, software fault perdition with the help of the Quadratic Censored regressive convolution deep neural network-based classification. The deep learning classifier analyzes the training and testing samples using the contingency correlation coefficient. The softstep activation function is used to provide the final fault prediction results. To minimize the error, the Nelder–Mead method is applied to solve non-linear least-squares problems. Finally, accurate classification results with a minimum error are obtained at the output layer. Experimental evaluation is carried out with different quantitative metrics such as accuracy, precision, recall, F-measure, and time complexity. The analyzed results demonstrate the superior performance of our proposed SQADEN technique with maximum accuracy, sensitivity and specificity by 3%, 3%, 2% and 3% and minimum time and space by 13% and 15% when compared with the two state-of-the-art methods. 展开更多
关键词 Software defect prediction feature selection nonparametric statistical Torgerson-Gower scaling technique quadratic censored regressive convolution deep neural network softstep activation function nelder-mead method
下载PDF
Deep Learning for Wind Speed Forecasting Using Bi-LSTM with Selected Features 被引量:1
5
作者 Siva Sankari Subbiah Senthil Kumar Paramasivan +2 位作者 Karmel Arockiasamy Saminathan Senthivel Muthamilselvan Thangavel 《Intelligent Automation & Soft Computing》 SCIE 2023年第3期3829-3844,共16页
Wind speed forecasting is important for wind energy forecasting.In the modern era,the increase in energy demand can be managed effectively by fore-casting the wind speed accurately.The main objective of this research ... Wind speed forecasting is important for wind energy forecasting.In the modern era,the increase in energy demand can be managed effectively by fore-casting the wind speed accurately.The main objective of this research is to improve the performance of wind speed forecasting by handling uncertainty,the curse of dimensionality,overfitting and non-linearity issues.The curse of dimensionality and overfitting issues are handled by using Boruta feature selec-tion.The uncertainty and the non-linearity issues are addressed by using the deep learning based Bi-directional Long Short Term Memory(Bi-LSTM).In this paper,Bi-LSTM with Boruta feature selection named BFS-Bi-LSTM is proposed to improve the performance of wind speed forecasting.The model identifies relevant features for wind speed forecasting from the meteorological features using Boruta wrapper feature selection(BFS).Followed by Bi-LSTM predicts the wind speed by considering the wind speed from the past and future time steps.The proposed BFS-Bi-LSTM model is compared against Multilayer perceptron(MLP),MLP with Boruta(BFS-MLP),Long Short Term Memory(LSTM),LSTM with Boruta(BFS-LSTM)and Bi-LSTM in terms of Root Mean Square Error(RMSE),Mean Absolute Error(MAE),Mean Square Error(MSE)and R2.The BFS-Bi-LSTM surpassed other models by producing RMSE of 0.784,MAE of 0.530,MSE of 0.615 and R2 of 0.8766.The experimental result shows that the BFS-Bi-LSTM produced better forecasting results compared to others. 展开更多
关键词 Bi-directional long short term memory boruta feature selection deep learning machine learning wind speed forecasting
下载PDF
A Framework of Deep Optimal Features Selection for Apple Leaf Diseases Recognition
6
作者 Samra Rehman Muhammad Attique Khan +5 位作者 Majed Alhaisoni Ammar Armghan Usman Tariq Fayadh Alenezi Ye Jin Kim Byoungchol Chang 《Computers, Materials & Continua》 SCIE EI 2023年第4期697-714,共18页
Identifying fruit disease manually is time-consuming, expertrequired,and expensive;thus, a computer-based automated system is widelyrequired. Fruit diseases affect not only the quality but also the quantity.As a resul... Identifying fruit disease manually is time-consuming, expertrequired,and expensive;thus, a computer-based automated system is widelyrequired. Fruit diseases affect not only the quality but also the quantity.As a result, it is possible to detect the disease early on and cure the fruitsusing computer-based techniques. However, computer-based methods faceseveral challenges, including low contrast, a lack of dataset for training amodel, and inappropriate feature extraction for final classification. In thispaper, we proposed an automated framework for detecting apple fruit leafdiseases usingCNNand a hybrid optimization algorithm. Data augmentationis performed initially to balance the selected apple dataset. After that, twopre-trained deep models are fine-tuning and trained using transfer learning.Then, a fusion technique is proposed named Parallel Correlation Threshold(PCT). The fused feature vector is optimized in the next step using a hybridoptimization algorithm. The selected features are finally classified usingmachine learning algorithms. Four different experiments have been carriedout on the augmented Plant Village dataset and yielded the best accuracy of99.8%. The accuracy of the proposed framework is also compared to that ofseveral neural nets, and it outperforms them all. 展开更多
关键词 Convolutional neural networks deep learning features fusion features optimization CLASSIFICATION
下载PDF
HRNetO:Human Action Recognition Using Unified Deep Features Optimization Framework
7
作者 Tehseen Ahsan Sohail Khalid +3 位作者 Shaheryar Najam Muhammad Attique Khan Ye Jin Kim Byoungchol Chang 《Computers, Materials & Continua》 SCIE EI 2023年第4期1089-1105,共17页
Human action recognition(HAR)attempts to understand a subject’sbehavior and assign a label to each action performed.It is more appealingbecause it has a wide range of applications in computer vision,such asvideo surv... Human action recognition(HAR)attempts to understand a subject’sbehavior and assign a label to each action performed.It is more appealingbecause it has a wide range of applications in computer vision,such asvideo surveillance and smart cities.Many attempts have been made in theliterature to develop an effective and robust framework for HAR.Still,theprocess remains difficult and may result in reduced accuracy due to severalchallenges,such as similarity among actions,extraction of essential features,and reduction of irrelevant features.In this work,we proposed an end-toendframework using deep learning and an improved tree seed optimizationalgorithm for accurate HAR.The proposed design consists of a fewsignificantsteps.In the first step,frame preprocessing is performed.In the second step,two pre-trained deep learning models are fine-tuned and trained throughdeep transfer learning using preprocessed video frames.In the next step,deeplearning features of both fine-tuned models are fused using a new ParallelStandard Deviation Padding Max Value approach.The fused features arefurther optimized using an improved tree seed algorithm,and select the bestfeatures are finally classified by using the machine learning classifiers.Theexperiment was carried out on five publicly available datasets,including UTInteraction,Weizmann,KTH,Hollywood,and IXAMS,and achieved higheraccuracy than previous techniques. 展开更多
关键词 Action recognition features fusion deep learning features selection
下载PDF
Human Gait Recognition Based on Sequential Deep Learning and Best Features Selection
8
作者 Ch Avais Hanif Muhammad Ali Mughal +3 位作者 Muhammad Attique Khan Usman Tariq Ye Jin Kim Jae-Hyuk Cha 《Computers, Materials & Continua》 SCIE EI 2023年第6期5123-5140,共18页
Gait recognition is an active research area that uses a walking theme to identify the subject correctly.Human Gait Recognition(HGR)is performed without any cooperation from the individual.However,in practice,it remain... Gait recognition is an active research area that uses a walking theme to identify the subject correctly.Human Gait Recognition(HGR)is performed without any cooperation from the individual.However,in practice,it remains a challenging task under diverse walking sequences due to the covariant factors such as normal walking and walking with wearing a coat.Researchers,over the years,have worked on successfully identifying subjects using different techniques,but there is still room for improvement in accuracy due to these covariant factors.This paper proposes an automated model-free framework for human gait recognition in this article.There are a few critical steps in the proposed method.Firstly,optical flow-based motion region esti-mation and dynamic coordinates-based cropping are performed.The second step involves training a fine-tuned pre-trained MobileNetV2 model on both original and optical flow cropped frames;the training has been conducted using static hyperparameters.The third step proposed a fusion technique known as normal distribution serially fusion.In the fourth step,a better optimization algorithm is applied to select the best features,which are then classified using a Bi-Layered neural network.Three publicly available datasets,CASIA A,CASIA B,and CASIA C,were used in the experimental process and obtained average accuracies of 99.6%,91.6%,and 95.02%,respectively.The proposed framework has achieved improved accuracy compared to the other methods. 展开更多
关键词 Human gait recognition optical flow deep learning features FUSION feature selection
下载PDF
GaitDONet: Gait Recognition Using Deep Features Optimization and Neural Network
9
作者 Muhammad Attique Khan Awais Khan +6 位作者 Majed Alhaisoni Abdullah Alqahtani Ammar Armghan Sara A.Althubiti Fayadh Alenezi Senghour Mey Yunyoung Nam 《Computers, Materials & Continua》 SCIE EI 2023年第6期5087-5103,共17页
Human gait recognition(HGR)is the process of identifying a sub-ject(human)based on their walking pattern.Each subject is a unique walking pattern and cannot be simulated by other subjects.But,gait recognition is not e... Human gait recognition(HGR)is the process of identifying a sub-ject(human)based on their walking pattern.Each subject is a unique walking pattern and cannot be simulated by other subjects.But,gait recognition is not easy and makes the system difficult if any object is carried by a subject,such as a bag or coat.This article proposes an automated architecture based on deep features optimization for HGR.To our knowledge,it is the first architecture in which features are fused using multiset canonical correlation analysis(MCCA).In the proposed method,original video frames are processed for all 11 selected angles of the CASIA B dataset and utilized to train two fine-tuned deep learning models such as Squeezenet and Efficientnet.Deep transfer learning was used to train both fine-tuned models on selected angles,yielding two new targeted models that were later used for feature engineering.Features are extracted from the deep layer of both fine-tuned models and fused into one vector using MCCA.An improved manta ray foraging optimization algorithm is also proposed to select the best features from the fused feature matrix and classified using a narrow neural network classifier.The experimental process was conducted on all 11 angles of the large multi-view gait dataset(CASIA B)dataset and obtained improved accuracy than the state-of-the-art techniques.Moreover,a detailed confidence interval based analysis also shows the effectiveness of the proposed architecture for HGR. 展开更多
关键词 Human gait recognition BIOMETRIC deep learning features fusion OPTIMIZATION neural network
下载PDF
Deep-Net:Fine-Tuned Deep Neural Network Multi-Features Fusion for Brain Tumor Recognition
10
作者 Muhammad Attique Khan Reham R.Mostafa +6 位作者 Yu-Dong Zhang Jamel Baili Majed Alhaisoni Usman Tariq Junaid Ali Khan Ye Jin Kim Jaehyuk Cha 《Computers, Materials & Continua》 SCIE EI 2023年第9期3029-3047,共19页
Manual diagnosis of brain tumors usingmagnetic resonance images(MRI)is a hectic process and time-consuming.Also,it always requires an expert person for the diagnosis.Therefore,many computer-controlled methods for diag... Manual diagnosis of brain tumors usingmagnetic resonance images(MRI)is a hectic process and time-consuming.Also,it always requires an expert person for the diagnosis.Therefore,many computer-controlled methods for diagnosing and classifying brain tumors have been introduced in the literature.This paper proposes a novel multimodal brain tumor classification framework based on two-way deep learning feature extraction and a hybrid feature optimization algorithm.NasNet-Mobile,a pre-trained deep learning model,has been fine-tuned and twoway trained on original and enhancedMRI images.The haze-convolutional neural network(haze-CNN)approach is developed and employed on the original images for contrast enhancement.Next,transfer learning(TL)is utilized for training two-way fine-tuned models and extracting feature vectors from the global average pooling layer.Then,using a multiset canonical correlation analysis(CCA)method,features of both deep learning models are fused into a single feature matrix—this technique aims to enhance the information in terms of features for better classification.Although the information was increased,computational time also jumped.This issue is resolved using a hybrid feature optimization algorithm that chooses the best classification features.The experiments were done on two publicly available datasets—BraTs2018 and BraTs2019—and yielded accuracy rates of 94.8%and 95.7%,respectively.The proposedmethod is comparedwith several recent studies andoutperformed inaccuracy.In addition,we analyze the performance of each middle step of the proposed approach and find the selection technique strengthens the proposed framework. 展开更多
关键词 Brain tumor haze contrast enhancement deep learning transfer learning features optimization
下载PDF
An Automated Classification Technique for COVID-19 Using Optimized Deep Learning Features
11
作者 Ejaz Khan Muhammad Zia Ur Rehman +3 位作者 Fawad Ahmed Suliman A.Alsuhibany Muhammad Zulfiqar Ali Jawad Ahmad 《Computer Systems Science & Engineering》 SCIE EI 2023年第9期3799-3814,共16页
In 2020,COVID-19 started spreading throughout the world.This deadly infection was identified as a virus that may affect the lungs and,in severe cases,could be the cause of death.The polymerase chain reaction(PCR)test ... In 2020,COVID-19 started spreading throughout the world.This deadly infection was identified as a virus that may affect the lungs and,in severe cases,could be the cause of death.The polymerase chain reaction(PCR)test is commonly used to detect this virus through the nasal passage or throat.However,the PCR test exposes health workers to this deadly virus.To limit human exposure while detecting COVID-19,image processing techniques using deep learning have been successfully applied.In this paper,a strategy based on deep learning is employed to classify the COVID-19 virus.To extract features,two deep learning models have been used,the DenseNet201 and the SqueezeNet.Transfer learning is used in feature extraction,and models are fine-tuned.A publicly available computerized tomography(CT)scan dataset has been used in this study.The extracted features from the deep learning models are optimized using the Ant Colony Optimization algorithm.The proposed technique is validated through multiple evaluation parameters.Several classifiers have been employed to classify the optimized features.The cubic support vector machine(Cubic SVM)classifier shows superiority over other commonly used classifiers and attained an accuracy of 98.72%.The proposed technique achieves state-of-the-art accuracy,a sensitivity of 98.80%,and a specificity of 96.64%. 展开更多
关键词 CT scans COVID-19 classification deep learning feature optimization
下载PDF
A Novel Handcrafted with Deep Features Based Brain Tumor Diagnosis Model
12
作者 Abdul Rahaman Wahab Sait Mohamad Khairi Ishak 《Intelligent Automation & Soft Computing》 SCIE 2023年第2期2057-2070,共14页
In healthcare sector,image classification is one of the crucial problems that impact the quality output from image processing domain.The purpose of image classification is to categorize different healthcare images under... In healthcare sector,image classification is one of the crucial problems that impact the quality output from image processing domain.The purpose of image classification is to categorize different healthcare images under various class labels which in turn helps in the detection and management of diseases.Magnetic Resonance Imaging(MRI)is one of the effective non-invasive strate-gies that generate a huge and distinct number of tissue contrasts in every imaging modality.This technique is commonly utilized by healthcare professionals for Brain Tumor(BT)diagnosis.With recent advancements in Machine Learning(ML)and Deep Learning(DL)models,it is possible to detect the tumor from images automatically,using a computer-aided design.The current study focuses on the design of automated Deep Learning-based BT Detection and Classification model using MRI images(DLBTDC-MRI).The proposed DLBTDC-MRI techni-que aims at detecting and classifying different stages of BT.The proposed DLBTDC-MRI technique involves medianfiltering technique to remove the noise and enhance the quality of MRI images.Besides,morphological operations-based image segmentation approach is also applied to determine the BT-affected regions in brain MRI image.Moreover,a fusion of handcrafted deep features using VGGNet is utilized to derive a valuable set of feature vectors.Finally,Artificial Fish Swarm Optimization(AFSO)with Artificial Neural Network(ANN)model is utilized as a classifier to decide the presence of BT.In order to assess the enhanced BT classification performance of the proposed model,a comprehensive set of simulations was performed on benchmark dataset and the results were vali-dated under several measures. 展开更多
关键词 Brain tumor medical imaging image classification handcrafted features deep learning parameter optimization
下载PDF
Deepfake Video Detection Employing Human Facial Features
13
作者 Daniel Schilling Weiss Nguyen Desmond T. Ademiluyi 《Journal of Computer and Communications》 2023年第12期1-13,共13页
Deepfake technology can be used to replace people’s faces in videos or pictures to show them saying or doing things they never said or did. Deepfake media are often used to extort, defame, and manipulate public opini... Deepfake technology can be used to replace people’s faces in videos or pictures to show them saying or doing things they never said or did. Deepfake media are often used to extort, defame, and manipulate public opinion. However, despite deepfake technology’s risks, current deepfake detection methods lack generalization and are inconsistent when applied to unknown videos, i.e., videos on which they have not been trained. The purpose of this study is to develop a generalizable deepfake detection model by training convoluted neural networks (CNNs) to classify human facial features in videos. The study formulated the research questions: “How effectively does the developed model provide reliable generalizations?” A CNN model was trained to distinguish between real and fake videos using the facial features of human subjects in videos. The model was trained, validated, and tested using the FaceForensiq++ dataset, which contains more than 500,000 frames and subsets of the DFDC dataset, totaling more than 22,000 videos. The study demonstrated high generalizability, as the accuracy of the unknown dataset was only marginally (about 1%) lower than that of the known dataset. The findings of this study indicate that detection systems can be more generalizable, lighter, and faster by focusing on just a small region (the human face) of an entire video. 展开更多
关键词 Artificial Intelligence Convoluted Neural Networks deepfake GANs GENERALIZATION deep Learning Facial features Video Frames
下载PDF
STGSA:A Novel Spatial-Temporal Graph Synchronous Aggregation Model for Traffic Prediction 被引量:1
14
作者 Zebing Wei Hongxia Zhao +5 位作者 Zhishuai Li Xiaojie Bu Yuanyuan Chen Xiqiao Zhang Yisheng Lv Fei-Yue Wang 《IEEE/CAA Journal of Automatica Sinica》 SCIE EI CSCD 2023年第1期226-238,共13页
The success of intelligent transportation systems relies heavily on accurate traffic prediction,in which how to model the underlying spatial-temporal information from traffic data has come under the spotlight.Most exi... The success of intelligent transportation systems relies heavily on accurate traffic prediction,in which how to model the underlying spatial-temporal information from traffic data has come under the spotlight.Most existing frameworks typically utilize separate modules for spatial and temporal correlations modeling.However,this stepwise pattern may limit the effectiveness and efficiency in spatial-temporal feature extraction and cause the overlook of important information in some steps.Furthermore,it is lacking sufficient guidance from prior information while modeling based on a given spatial adjacency graph(e.g.,deriving from the geodesic distance or approximate connectivity),and may not reflect the actual interaction between nodes.To overcome those limitations,our paper proposes a spatial-temporal graph synchronous aggregation(STGSA)model to extract the localized and long-term spatial-temporal dependencies simultaneously.Specifically,a tailored graph aggregation method in the vertex domain is designed to extract spatial and temporal features in one graph convolution process.In each STGSA block,we devise a directed temporal correlation graph to represent the localized and long-term dependencies between nodes,and the potential temporal dependence is further fine-tuned by an adaptive weighting operation.Meanwhile,we construct an elaborated spatial adjacency matrix to represent the road sensor graph by considering both physical distance and node similarity in a datadriven manner.Then,inspired by the multi-head attention mechanism which can jointly emphasize information from different r epresentation subspaces,we construct a multi-stream module based on the STGSA blocks to capture global information.It projects the embedding input repeatedly with multiple different channels.Finally,the predicted values are generated by stacking several multi-stream modules.Extensive experiments are constructed on six real-world datasets,and numerical results show that the proposed STGSA model significantly outperforms the benchmarks. 展开更多
关键词 deep learning graph neural network(GNN) multistream spatial-temporal feature extraction temporal graph traffic prediction
下载PDF
Soft Tissue Feature Tracking Based on Deep Matching Network
15
作者 Siyu Lu Shan Liu +4 位作者 Pengfei Hou Bo Yang Mingzhe Liu Lirong Yin Wenfeng Zheng 《Computer Modeling in Engineering & Sciences》 SCIE EI 2023年第7期363-379,共17页
Research in the field ofmedical image is an important part of themedical robot to operate human organs.Amedical robot is the intersection ofmulti-disciplinary research fields,in whichmedical image is an important dire... Research in the field ofmedical image is an important part of themedical robot to operate human organs.Amedical robot is the intersection ofmulti-disciplinary research fields,in whichmedical image is an important direction and has achieved fruitful results.In this paper,amethodof soft tissue surface feature tracking basedonadepthmatching network is proposed.This method is described based on the triangular matching algorithm.First,we construct a self-made sample set for training the depth matching network from the first N frames of speckle matching data obtained by the triangle matching algorithm.The depth matching network is pre-trained on the ORL face data set and then trained on the self-made training set.After the training,the speckle matching is carried out in the subsequent frames to obtain the speckle matching matrix between the subsequent frames and the first frame.From this matrix,the inter-frame feature matching results can be obtained.In this way,the inter-frame speckle tracking is completed.On this basis,the results of this method are compared with the matching results based on the convolutional neural network.The experimental results show that the proposed method has higher matching accuracy.In particular,the accuracy of the MNIST handwritten data set has reached more than 90%. 展开更多
关键词 Soft tissue feature tracking deep matching network
下载PDF
Feature Selection with Deep Belief Network for Epileptic Seizure Detection on EEG Signals
16
作者 Srikanth Cherukuvada R.Kayalvizhi 《Computers, Materials & Continua》 SCIE EI 2023年第5期4101-4118,共18页
The term Epilepsy refers to a most commonly occurring brain disorder after a migraine.Early identification of incoming seizures significantly impacts the lives of people with Epilepsy.Automated detection of epileptic ... The term Epilepsy refers to a most commonly occurring brain disorder after a migraine.Early identification of incoming seizures significantly impacts the lives of people with Epilepsy.Automated detection of epileptic seizures(ES)has dramatically improved the life quality of the patients.Recent Electroencephalogram(EEG)related seizure detection mechanisms encountered several difficulties in real-time.The EEGs are the non-stationary signal,and seizure patternswould changewith patients and recording sessions.Further,EEG data were disposed to wide noise varieties that adversely moved the recognition accuracy of ESs.Artificial intelligence(AI)methods in the domain of ES analysis use traditional deep learning(DL),and machine learning(ML)approaches.This article introduces an Oppositional Aquila Optimizer-based Feature Selection with Deep Belief Network for Epileptic Seizure Detection(OAOFS-DBNECD)technique using EEG signals.The primary aim of the presented OAOFS-DBNECD system is to categorize and classify the presence of ESs.The suggested OAOFS-DBNECD technique transforms the EEG signals into.csv format at the initial stage.Next,the OAOFS technique selects an optimal subset of features using the preprocessed data.For seizure classification,the presented OAOFS-DBNECD technique applies Artificial Ecosystem Optimizer(AEO)with a deep belief network(DBN)model.An extensive range of simulations was performed on the benchmark dataset to ensure the enhanced performance of the presented OAOFS-DBNECD algorithm.The comparison study shows the significant outcomes of the OAOFS-DBNECD approach over other methodologies.In addition,the result of the suggested approach has been evaluated using the CHB-MIT database,and the findings demonstrate accuracy of 97.81%.These findings confirmed the best seizure categorization accuracy on the EEG data considered. 展开更多
关键词 Seizure detection EEG signals machine learning deep learning feature selection
下载PDF
A Fusion of Residual Blocks and Stack Auto Encoder Features for Stomach Cancer Classification
17
作者 Abdul Haseeb Muhammad Attique Khan +5 位作者 Majed Alhaisoni Ghadah Aldehim Leila Jamel Usman Tariq Taerang Kim Jae-Hyuk Cha 《Computers, Materials & Continua》 SCIE EI 2023年第12期3895-3920,共26页
Diagnosing gastrointestinal cancer by classical means is a hazardous procedure.Years have witnessed several computerized solutions for stomach disease detection and classification.However,the existing techniques faced... Diagnosing gastrointestinal cancer by classical means is a hazardous procedure.Years have witnessed several computerized solutions for stomach disease detection and classification.However,the existing techniques faced challenges,such as irrelevant feature extraction,high similarity among different disease symptoms,and the least-important features from a single source.This paper designed a new deep learning-based architecture based on the fusion of two models,Residual blocks and Auto Encoder.First,the Hyper-Kvasir dataset was employed to evaluate the proposed work.The research selected a pre-trained convolutional neural network(CNN)model and improved it with several residual blocks.This process aims to improve the learning capability of deep models and lessen the number of parameters.Besides,this article designed an Auto-Encoder-based network consisting of five convolutional layers in the encoder stage and five in the decoder phase.The research selected the global average pooling and convolutional layers for the feature extraction optimized by a hybrid Marine Predator optimization and Slime Mould optimization algorithm.These features of both models are fused using a novel fusion technique that is later classified using the Artificial Neural Network classifier.The experiment worked on the HyperKvasir dataset,which consists of 23 stomach-infected classes.At last,the proposed method obtained an improved accuracy of 93.90%on this dataset.Comparison is also conducted with some recent techniques and shows that the proposed method’s accuracy is improved. 展开更多
关键词 Gastrointestinal cancer contrast enhancement deep learning information fusion feature selection machine learning
下载PDF
A Novel Efficient Patient Monitoring FER System Using Optimal DL-Features
18
作者 Mousa Alhajlah 《Computers, Materials & Continua》 SCIE EI 2023年第3期6161-6175,共15页
Automated Facial Expression Recognition(FER)serves as the backbone of patient monitoring systems,security,and surveillance systems.Real-time FER is a challenging task,due to the uncontrolled nature of the environment ... Automated Facial Expression Recognition(FER)serves as the backbone of patient monitoring systems,security,and surveillance systems.Real-time FER is a challenging task,due to the uncontrolled nature of the environment and poor quality of input frames.In this paper,a novel FER framework has been proposed for patient monitoring.Preprocessing is performed using contrast-limited adaptive enhancement and the dataset is balanced using augmentation.Two lightweight efficient Convolution Neural Network(CNN)models MobileNetV2 and Neural search Architecture Network Mobile(NasNetMobile)are trained,and feature vectors are extracted.The Whale Optimization Algorithm(WOA)is utilized to remove irrelevant features from these vectors.Finally,the optimized features are serially fused to pass them to the classifier.A comprehensive set of experiments were carried out for the evaluation of real-time image datasets FER-2013,MMA,and CK+to report performance based on various metrics.Accuracy results show that the proposed model has achieved 82.5%accuracy and performed better in comparison to the state-of-the-art classification techniques in terms of accuracy.We would like to highlight that the proposed technique has achieved better accuracy by using 2.8 times lesser number of features. 展开更多
关键词 Facial expression recognition deep learning transfer learning feature optimization
下载PDF
Basic Features of the Crustal Structure in the Lower Yangtze and Its Neighboring Area in the Chinese Mainland: Review of Deep Seismic Sounding Research 被引量:3
19
作者 Bai Zhiming Wu Qingju +1 位作者 Xu Tao Wang Xiao 《Earthquake Research in China》 CSCD 2016年第3期298-315,共18页
The Deep Seismic Sounding( DSS) projects carried out from the 1970 s in the lower Yangtze region and its neighboring area were reviewed in this paper,then the basic wave group features of those wide angle reflection /... The Deep Seismic Sounding( DSS) projects carried out from the 1970 s in the lower Yangtze region and its neighboring area were reviewed in this paper,then the basic wave group features of those wide angle reflection / refraction record sections,and of the crustal structure are summarized. It shows that there were in total five clear wave groups on the record sections,which include the first arrival Pg,the reflection P1 from the bottom interface of the upper crust,the reflection P3 from the bottom interface of the middle crust,the strong reflection Pm from the Moho boundary,and the refraction Pn from uppermost mantle. In general,these phases are easily consistently traced and compared,despite some first arrivals being delayed or arriving earlier than normal due to the shallow sedimentary cover or bedrocks. In particular,in the Dabie Mountain region the seismic events of a few gathered shots always have weak reflection energy,are twisted,or exhibit disorganized waveforms, which could be attributed to the disruption variations of reflection depth,the broken Moho,and the discontinuity of the reflection boundary within crust. The regional crustal structures are composed of the upper,middle and lower crust,of which the middle and lower layers can be divided into two weak reflection ones. The crustal thickness of the North China and Yangtze platform are 30km- 36 km,and the Moho exhibits a flat geometry despite some local uplifts. The average pressure velocity in lower crust beneath this two tectonic area is 6. 7 ± 0. 3km / s. Nevertheless,beneath the Dabieshan area the crustal thickness is 32km- 41 km,the Moho bends down sharply andtakes an abrupt 4km- 7km dislocation in the vertical direction. The average pressure velocity in the lower crust beneath the Dabieshan area is 6. 8 ± 0. 2km / s. 展开更多
关键词 deep seismic sounding projects Wave group feature Crustal structure feature Anhui Province and its neighboring area
下载PDF
Automated Pavement Crack Detection Using Deep Feature Selection and Whale Optimization Algorithm
20
作者 Shorouq Alshawabkeh Li Wu +3 位作者 Daojun Dong Yao Cheng Liping Li Mohammad Alanaqreh 《Computers, Materials & Continua》 SCIE EI 2023年第10期63-77,共15页
Pavement crack detection plays a crucial role in ensuring road safety and reducing maintenance expenses.Recent advancements in deep learning(DL)techniques have shown promising results in detecting pavement cracks;howe... Pavement crack detection plays a crucial role in ensuring road safety and reducing maintenance expenses.Recent advancements in deep learning(DL)techniques have shown promising results in detecting pavement cracks;however,the selection of relevant features for classification remains challenging.In this study,we propose a new approach for pavement crack detection that integrates deep learning for feature extraction,the whale optimization algorithm(WOA)for feature selection,and random forest(RF)for classification.The performance of the models was evaluated using accuracy,recall,precision,F1 score,and area under the receiver operating characteristic curve(AUC).Our findings reveal that Model 2,which incorporates RF into the ResNet-18 architecture,outperforms baseline Model 1 across all evaluation metrics.Nevertheless,our proposed model,which combines ResNet-18 with both WOA and RF,achieves significantly higher accuracy,recall,precision,and F1 score compared to the other two models.These results underscore the effectiveness of integrating RF and WOA into ResNet-18 for pavement crack detection applications.We applied the proposed approach to a dataset of pavement images,achieving an accuracy of 97.16%and an AUC of 0.984.Our results demonstrate that the proposed approach surpasses existing methods for pavement crack detection,offering a promising solution for the automatic identification of pavement cracks.By leveraging this approach,potential safety hazards can be identified more effectively,enabling timely repairs and maintenance measures.Lastly,the findings of this study also emphasize the potential of integrating RF and WOA with deep learning for pavement crack detection,providing road authorities with the necessary tools to make informed decisions regarding road infrastructure maintenance. 展开更多
关键词 Pavement crack detection deep learning feature selection whale optimization algorithm civil engineering
下载PDF
上一页 1 2 213 下一页 到第
使用帮助 返回顶部