Early detection of the Covid-19 disease is essential due to its higher rate of infection affecting tens of millions of people,and its high number of deaths also by 7%.For that purpose,a proposed model of several stage...Early detection of the Covid-19 disease is essential due to its higher rate of infection affecting tens of millions of people,and its high number of deaths also by 7%.For that purpose,a proposed model of several stages was developed.The first stage is optimizing the images using dynamic adaptive histogram equalization,performing a semantic segmentation using DeepLabv3Plus,then augmenting the data by flipping it horizontally,rotating it,then flipping it vertically.The second stage builds a custom convolutional neural network model using several pre-trained ImageNet.Finally,the model compares the pre-trained data to the new output,while repeatedly trimming the best-performing models to reduce complexity and improve memory efficiency.Several experiments were done using different techniques and parameters.Accordingly,the proposed model achieved an average accuracy of 99.6%and an area under the curve of 0.996 in the Covid-19 detection.This paper will discuss how to train a customized intelligent convolutional neural network using various parameters on a set of chest X-rays with an accuracy of 99.6%.展开更多
The Wavelet-Domain Projection Pursuit Learning Network (WDPPLN) is proposedfor restoring degraded image. The new network combines the advantages of both projectionpursuit and wavelet shrinkage. Restoring image is very...The Wavelet-Domain Projection Pursuit Learning Network (WDPPLN) is proposedfor restoring degraded image. The new network combines the advantages of both projectionpursuit and wavelet shrinkage. Restoring image is very difficult when little is known about apriori knowledge for multisource degraded factors. WDPPLN successfully resolves this problemby separately processing wavelet coefficients and scale coefficients. Parameters in WDPPLN,which are used to simulate degraded factors, are estimated via WDPPLN training, using scalecoefficients. Also, WDPPLN uses soft-threshold of wavelet shrinkage technique to suppress noisein three high frequency subbands. The new method is compared with the traditional methodsand the Projection Pursuit Learning Network (PPLN) method. Experimental results demonstratethat it is an effective method for unsupervised restoring degraded image.展开更多
In the electricity market,fluctuations in real-time prices are unstable,and changes in short-term load are determined by many factors.By studying the timing of charging and discharging,as well as the economic benefits...In the electricity market,fluctuations in real-time prices are unstable,and changes in short-term load are determined by many factors.By studying the timing of charging and discharging,as well as the economic benefits of energy storage in the process of participating in the power market,this paper takes energy storage scheduling as merely one factor affecting short-term power load,which affects short-term load time series along with time-of-use price,holidays,and temperature.A deep learning network is used to predict the short-term load,a convolutional neural network(CNN)is used to extract the features,and a long short-term memory(LSTM)network is used to learn the temporal characteristics of the load value,which can effectively improve prediction accuracy.Taking the load data of a certain region as an example,the CNN-LSTM prediction model is compared with the single LSTM prediction model.The experimental results show that the CNN-LSTM deep learning network with the participation of energy storage in dispatching can have high prediction accuracy for short-term power load forecasting.展开更多
Landfill leaks pose a serious threat to environmental health,risking the contamination of both groundwater and soil resources.Accurate investigation of these sites is essential for implementing effective prevention an...Landfill leaks pose a serious threat to environmental health,risking the contamination of both groundwater and soil resources.Accurate investigation of these sites is essential for implementing effective prevention and control measures.The self-potential(SP)stands out for its sensitivity to contamination plumes,offering a solution for monitoring and detecting the movement and seepage of subsurface pollutants.However,traditional SP inversion techniques heavily rely on precise subsurface resistivity information.In this study,we propose the Attention U-Net deep learning network for rapid SP inversion.By incorporating an attention mechanism,this algorithm effectively learns the relationship between array-style SP data and the location and extent of subsurface contaminated sources.We designed a synthetic landfill model with a heterogeneous resistivity structure to assess the performance of Attention U-Net deep learning network.Additionally,we conducted further validation using a laboratory model to assess its practical applicability.The results demonstrate that the algorithm is not solely dependent on resistivity information,enabling effective locating of the source distribution,even in models with intricate subsurface structures.Our work provides a promising tool for SP data processing,enhancing the applicability of this method in the field of near-subsurface environmental monitoring.展开更多
This study investigated how students used peer assessments in synchronous learning network (SLN) to assess each other s writing. It focused on examining the frequency and styles of various techniques students employed...This study investigated how students used peer assessments in synchronous learning network (SLN) to assess each other s writing. It focused on examining the frequency and styles of various techniques students employed while assessing each others writing and student response to assessing each other s writing in a SLN context. The findings indicated that these students received many assessments during each peer assessment activity. They preferred to use assessing techniques of less critical types, and had po...展开更多
The transparent open box(TOB)learning network algorithm offers an alternative approach to the lack of transparency provided by most machine-learning algorithms.It provides the exact calculations and relationships amon...The transparent open box(TOB)learning network algorithm offers an alternative approach to the lack of transparency provided by most machine-learning algorithms.It provides the exact calculations and relationships among the underlying input variables of the datasets to which it is applied.It also has the capability to achieve credible and auditable levels of prediction accuracy to complex,non-linear datasets,typical of those encountered in the oil and gas sector,highlighting the potential for underfitting and overfitting.The algorithm is applied here to predict bubble-point pressure from a published PVT dataset of 166 data records involving four easy-tomeasure variables(reservoir temperature,gas-oil ratio,oil gravity,gas density relative to air)with uneven,and in parts,sparse data coverage.The TOB network demonstrates high-prediction accuracy for this complex system,although it predictions applied to the full dataset are outperformed by an artificial neural network(ANN).However,the performance of the TOB algorithm reveals the risk of overfitting in the sparse areas of the dataset and achieves a prediction performance that matches the ANN algorithm where the underlying data population is adequate.The high levels of transparency and its inhibitions to overfitting enable the TOB learning network to provide complementary information about the underlying dataset to that provided by traditional machine learning algorithms.This makes them suitable for application in parallel with neural-network algorithms,to overcome their black-box tendencies,and for benchmarking the prediction performance of other machine learning algorithms.展开更多
Purpose-Parkinson’s disease(PD)is a well-known complex neurodegenerative disease.Typically,its identification is based on motor disorders,while the computer estimation of its main symptoms with computational machine ...Purpose-Parkinson’s disease(PD)is a well-known complex neurodegenerative disease.Typically,its identification is based on motor disorders,while the computer estimation of its main symptoms with computational machine learning(ML)has a high exposure which is supported by researches conducted.Nevertheless,ML approaches required first to refine their parameters and then to work with the best model generated.This process often requires an expert user to oversee the performance of the algorithm.Therefore,an attention is required towards new approaches for better forecasting accuracy.Design/methodology/approach-To provide an available identification model for Parkinson disease as an auxiliary function for clinicians,the authors suggest a new evolutionary classification model.The core of the prediction model is a fast learning network(FLN)optimized by a genetic algorithm(GA).To get a better subset of features and parameters,a new coding architecture is introduced to improve GA for obtaining an optimal FLN model.Findings-The proposed model is intensively evaluated through a series of experiments based on Speech and HandPD benchmark datasets.The very popular wrappers induction models such as support vector machine(SVM),K-nearest neighbors(KNN)have been tested in the same condition.The results support that the proposed model can achieve the best performances in terms of accuracy and g-mean.Originality/value-A novel efficient PD detectionmodel is proposed,which is called A-W-FLN.The A-W-FLN utilizes FLN as the base classifier;in order to take its higher generalization ability,and identification capability is alsoembedded to discover themost suitable featuremodel in the detection process.Moreover,the proposedmethod automatically optimizes the FLN’s architecture to a smaller number of hidden nodes and solid connecting weights.This helps the network to train on complex PD datasets with non-linear features and yields superior result.展开更多
With an increasing number of services connected to the internet,including cloud computing and Internet of Things(IoT)systems,the prevention of cyberattacks has become more challenging due to the high dimensionality of...With an increasing number of services connected to the internet,including cloud computing and Internet of Things(IoT)systems,the prevention of cyberattacks has become more challenging due to the high dimensionality of the network traffic data and access points.Recently,researchers have suggested deep learning(DL)algorithms to define intrusion features through training empirical data and learning anomaly patterns of attacks.However,due to the high dynamics and imbalanced nature of the data,the existing DL classifiers are not completely effective at distinguishing between abnormal and normal behavior line connections for modern networks.Therefore,it is important to design a self-adaptive model for an intrusion detection system(IDS)to improve the detection of attacks.Consequently,in this paper,a novel hybrid weighted deep belief network(HW-DBN)algorithm is proposed for building an efficient and reliable IDS(DeepIoT.IDS)model to detect existing and novel cyberattacks.The HW-DBN algorithm integrates an improved Gaussian–Bernoulli restricted Boltzmann machine(Deep GB-RBM)feature learning operator with a weighted deep neural networks(WDNN)classifier.The CICIDS2017 dataset is selected to evaluate the DeepIoT.IDS model as it contains multiple types of attacks,complex data patterns,noise values,and imbalanced classes.We have compared the performance of the DeepIoT.IDS model with three recent models.The results show the DeepIoT.IDS model outperforms the three other models by achieving a higher detection accuracy of 99.38%and 99.99%for web attack and bot attack scenarios,respectively.Furthermore,it can detect the occurrence of low-frequency attacks that are undetectable by other models.展开更多
Structure learning of Bayesian networks is a wellresearched but computationally hard task.For learning Bayesian networks,this paper proposes an improved algorithm based on unconstrained optimization and ant colony opt...Structure learning of Bayesian networks is a wellresearched but computationally hard task.For learning Bayesian networks,this paper proposes an improved algorithm based on unconstrained optimization and ant colony optimization(U-ACO-B) to solve the drawbacks of the ant colony optimization(ACO-B).In this algorithm,firstly,an unconstrained optimization problem is solved to obtain an undirected skeleton,and then the ACO algorithm is used to orientate the edges,thus returning the final structure.In the experimental part of the paper,we compare the performance of the proposed algorithm with ACO-B algorithm.The experimental results show that our method is effective and greatly enhance convergence speed than ACO-B algorithm.展开更多
To get simpler operation in modified fuzzy adaptive learning control network (FALCON) in some engineering application, sigmoid nonlinear function is employed as a substitute of traditional Gaussian membership functi...To get simpler operation in modified fuzzy adaptive learning control network (FALCON) in some engineering application, sigmoid nonlinear function is employed as a substitute of traditional Gaussian membership function. For making the modified FALCON learning more efficient and stable, a simulated annealing (SA) learning coefficient is introduced into learning algorithm. At first, the basic concepts and main advantages of FALCON were briefly reviewed. Subsequently, the topological structure and nodes operation were illustrated; the gradient-descent learning algorithm with SA learning coefficient was derived; and the distinctions between the archetype and the modification were analyzed. Eventually, the significance and worthiness of the modified FALCON were validated by its application to probability prediction of anode effect in aluminium electrolysis cells.展开更多
The homogeneity analysis of multi-airport system can provide important decision-making support for the route layout and cooperative operation.Existing research seldom analyzes the homogeneity of multi-airport system f...The homogeneity analysis of multi-airport system can provide important decision-making support for the route layout and cooperative operation.Existing research seldom analyzes the homogeneity of multi-airport system from the perspective of route network analysis,and the attribute information of airport nodes in the airport route network is not appropriately integrated into the airport network.In order to solve this problem,a multi-airport system homogeneity analysis method based on airport attribute network representation learning is proposed.Firstly,the route network of a multi-airport system with attribute information is constructed.If there are flights between airports,an edge is added between airports,and regional attribute information is added for each airport node.Secondly,the airport attributes and the airport network vector are represented respectively.The airport attributes and the airport network vector are embedded into the unified airport representation vector space by the network representation learning method,and then the airport vector integrating the airport attributes and the airport network characteristics is obtained.By calculating the similarity of the airport vectors,it is convenient to calculate the degree of homogeneity between airports and the homogeneity of the multi-airport system.The experimental results on the Beijing-Tianjin-Hebei multi-airport system show that,compared with other existing algorithms,the homogeneity analysis method based on attributed network representation learning can get more consistent results with the current situation of Beijing-Tianjin-Hebei multi-airport system.展开更多
Frequent counting is a very so often required operation in machine learning algorithms. A typical machine learning task, learning the structure of Bayesian network (BN) based on metric scoring, is introduced as an e...Frequent counting is a very so often required operation in machine learning algorithms. A typical machine learning task, learning the structure of Bayesian network (BN) based on metric scoring, is introduced as an example that heavily relies on frequent counting. A fast calculation method for frequent counting enhanced with two cache layers is then presented for learning BN. The main contribution of our approach is to eliminate comparison operations for frequent counting by introducing a multi-radix number system calculation. Both mathematical analysis and empirical comparison between our method and state-of-the-art solution are conducted. The results show that our method is dominantly superior to state-of-the-art solution in solving the problem of learning BN.展开更多
This paper first proposes a new self-learning data-driven methodology that can develop the failure criteria of unknown anisotropic ductile materials from the minimal number of experimental tests.Establishing failure c...This paper first proposes a new self-learning data-driven methodology that can develop the failure criteria of unknown anisotropic ductile materials from the minimal number of experimental tests.Establishing failure criteria of anisotropic ductile materials requires time-consuming tests and manual data evaluation.The proposed method can overcome such practical challenges.The methodology is formalized by combining four ideas:1)The deep learning neural network(DLNN)-based material constitutive model,2)Self-learning inverse finite element(SELIFE)simulation,3)Algorithmic identification of failure points from the selflearned stress-strain curves and 4)Derivation of the failure criteria through symbolic regression of the genetic programming.Stress update and the algorithmic tangent operator were formulated in terms of DLNN parameters for nonlinear finite element analysis.Then,the SELIFE simulation algorithm gradually makes the DLNN model learn highly complex multi-axial stress and strain relationships,being guided by the experimental boundary measurements.Following the failure point identification,a self-learning data-driven failure criteria are eventually developed with the help of a reliable symbolic regression algorithm.The methodology and the self-learning data-driven failure criteria were verified by comparing with a reference failure criteria and simulating with different materials orientations,respectively.展开更多
Many network presentation learning algorithms(NPLA)have originated from the process of the random walk between nodes in recent years.Despite these algorithms can obtain great embedding results,there may be also some l...Many network presentation learning algorithms(NPLA)have originated from the process of the random walk between nodes in recent years.Despite these algorithms can obtain great embedding results,there may be also some limitations.For instance,only the structural information of nodes is considered when these kinds of algorithms are constructed.Aiming at this issue,a label and community information-based network presentation learning algorithm(LC-NPLA)is proposed in this paper.First of all,by using the community information and the label information of nodes,the first-order neighbors of nodes are reconstructed.In the next,the random walk strategy is improved by integrating the degree information and label information of nodes.Then,the node sequence obtained from random walk sampling is transformed into the node representation vector by the Skip-Gram model.At last,the experimental results on ten real-world networks demonstrate that the proposed algorithm has great advantages in the label classification,network reconstruction and link prediction tasks,compared with three benchmark algorithms.展开更多
Conventional wing aerodynamic optimization processes can be time-consuming and imprecise due to the complexity of versatile flight missions.Plenty of existing literature has considered two-dimensional infinite airfoil...Conventional wing aerodynamic optimization processes can be time-consuming and imprecise due to the complexity of versatile flight missions.Plenty of existing literature has considered two-dimensional infinite airfoil optimization,while three-dimensional finite wing optimizations are subject to limited study because of high computational costs.Here we create an adaptive optimization methodology built upon digitized wing shape deformation and deep learning algorithms,which enable the rapid formulation of finite wing designs for specific aerodynamic performance demands under different cruise conditions.This methodology unfolds in three stages:radial basis function interpolated wing generation,collection of inputs from computational fluid dynamics simulations,and deep neural network that constructs the surrogate model for the optimal wing configuration.It has been demonstrated that the proposed methodology can significantly reduce the computational cost of numerical simulations.It also has the potential to optimize various aerial vehicles undergoing different mission environments,loading conditions,and safety requirements.展开更多
There are various heterogeneous networks for terminals to deliver a better quality of service. Signal system recognition and classification contribute a lot to the process. However, in low signal to noise ratio(SNR)...There are various heterogeneous networks for terminals to deliver a better quality of service. Signal system recognition and classification contribute a lot to the process. However, in low signal to noise ratio(SNR) circumstances or under time-varying multipath channels, the majority of the existing algorithms for signal recognition are already facing limitations. In this series, we present a robust signal recognition method based upon the original and latest updated version of the extreme learning machine(ELM) to help users to switch between networks. The ELM utilizes signal characteristics to distinguish systems. The superiority of this algorithm lies in the random choices of hidden nodes and in the fact that it determines the output weights analytically, which result in lower complexity. Theoretically, the algorithm tends to offer a good generalization performance at an extremely fast speed of learning. Moreover, we implement the GSM/WCDMA/LTE models in the Matlab environment by using the Simulink tools. The simulations reveal that the signals can be recognized successfully to achieve a 95% accuracy in a low SNR(0 dB) environment in the time-varying multipath Rayleigh fading channel.展开更多
Recommendation Information Systems(RIS)are pivotal in helping users in swiftly locating desired content from the vast amount of information available on the Internet.Graph Convolution Network(GCN)algorithms have been ...Recommendation Information Systems(RIS)are pivotal in helping users in swiftly locating desired content from the vast amount of information available on the Internet.Graph Convolution Network(GCN)algorithms have been employed to implement the RIS efficiently.However,the GCN algorithm faces limitations in terms of performance enhancement owing to the due to the embedding value-vanishing problem that occurs during the learning process.To address this issue,we propose a Weighted Forwarding method using the GCN(WF-GCN)algorithm.The proposed method involves multiplying the embedding results with different weights for each hop layer during graph learning.By applying the WF-GCN algorithm,which adjusts weights for each hop layer before forwarding to the next,nodes with many neighbors achieve higher embedding values.This approach facilitates the learning of more hop layers within the GCN framework.The efficacy of the WF-GCN was demonstrated through its application to various datasets.In the MovieLens dataset,the implementation of WF-GCN in LightGCN resulted in significant performance improvements,with recall and NDCG increasing by up to+163.64%and+132.04%,respectively.Similarly,in the Last.FM dataset,LightGCN using WF-GCN enhanced with WF-GCN showed substantial improvements,with the recall and NDCG metrics rising by up to+174.40%and+169.95%,respectively.Furthermore,the application of WF-GCN to Self-supervised Graph Learning(SGL)and Simple Graph Contrastive Learning(SimGCL)also demonstrated notable enhancements in both recall and NDCG across these datasets.展开更多
Computer-empowered detection of possible faults for Heating,Ventilation and Air-Conditioning(HVAC)subsystems,e.g.,chillers,is one of the most important applications in Artificial Intelligence(AI)integrated Internet of...Computer-empowered detection of possible faults for Heating,Ventilation and Air-Conditioning(HVAC)subsystems,e.g.,chillers,is one of the most important applications in Artificial Intelligence(AI)integrated Internet of Things(IoT).The cyber-physical system greatly enhances the safety and security of the working facilities,reducing time,saving energy and protecting humans’health.Under the current trends of smart building design and energy management optimization,Automated Fault Detection and Diagnosis(AFDD)of chillers integrated with IoT is highly demanded.Recent studies show that standard machine learning techniques,such as Principal Component Analysis(PCA),Support Vector Machine(SVM)and tree-structure-based algorithms,are useful in capturing various chiller faults with high accuracy rates.With the fast development of deep learning technology,Convolutional Neural Networks(CNNs)have been widely and successfully applied to various fields.However,for chiller AFDD,few existing works are adopting CNN and its extensions in the feature extraction and classification processes.In this study,we propose to perform chiller FDD using a CNN-based approach.The proposed approach has two distinct advantages over existing machine learning-based chiller AFDD methods.First,the CNN-based approach does not require the feature selection/extraction process.Since CNN is reputable with its feature extraction capability,the feature extraction and classification processes are merged,leading to a more neat AFDD framework compared to traditional approaches.Second,the classification accuracy is significantly improved compared to traditional methods using the CNN-based approach.展开更多
Flash floods are one of the most dangerous natural disasters,especially in hilly terrain,causing loss of life,property,and infrastructures and sudden disruption of traffic.These types of floods are mostly associated w...Flash floods are one of the most dangerous natural disasters,especially in hilly terrain,causing loss of life,property,and infrastructures and sudden disruption of traffic.These types of floods are mostly associated with landslides and erosion of roads within a short time.Most of Vietnamis hilly and mountainous;thus,the problem due to flash flood is severe and requires systematic studies to correctly identify flood susceptible areas for proper landuse planning and traffic management.In this study,three Machine Learning(ML)methods namely Deep Learning Neural Network(DL),Correlation-based FeatureWeighted Naive Bayes(CFWNB),and Adaboost(AB-CFWNB)were used for the development of flash flood susceptibility maps for hilly road section(115 km length)of National Highway(NH)-6 inHoa Binh province,Vietnam.In the proposedmodels,88 past flash flood events were used together with 14 flash floods affecting topographical and geo-environmental factors.The performance of themodels was evaluated using standard statisticalmeasures including Receiver Operating Characteristic(ROC)Curve,Area Under Curve(AUC)and Root Mean Square Error(RMSE).The results revealed that all the models performed well(AUC>0.80)in predicting flash flood susceptibility zones,but the performance of the DL model is the best(AUC:0.972,RMSE:0.352).Therefore,the DL model can be applied to develop an accurate flash flood susceptibility map of hilly terrain which can be used for proper planning and designing of the highways and other infrastructure facilities besides landuse management of the area.展开更多
Existing biomimetic robots can perform some basic rat-like movement primitives(MPs)and simple behavior with stiff combinations of these MPs.To mimic typical rat behavior with high similarity,we propose parameterizing ...Existing biomimetic robots can perform some basic rat-like movement primitives(MPs)and simple behavior with stiff combinations of these MPs.To mimic typical rat behavior with high similarity,we propose parameterizing the behavior using a probabilistic model and movement characteristics.First,an analysis of fifteen 10 min video sequences revealed that an actual rat has six typical behaviors in the open field,and each kind of behavior contains different bio-inspired combinations of eight MPs.We used the softmax classifier to obtain the behavior-movement hierarchical probability model.Secondly,we specified the MPs using movement parameters that are static and dynamic.We obtained the predominant values of the static and dynamic movement parameters using hierarchical clustering and fuzzy C-means clustering,respectively.These predominant parameters were used for fitting the rat spinal joint trajectory using a second-order Fourier series,and the joint trajectory was generalized using a back propagation neural network with two hidden layers.Finally,the hierarchical probability model and the generalized joint trajectory were mapped to the robot as control policy and commands,respectively.We implemented the six typical behaviors on the robot,and the results show high similarity when compared with the behaviors of actual rats.展开更多
基金This work was supported by the National Research Foundation of Korea-Grant funded by the Korean Government(Ministry of Science and ICT)-NRF-2020R1A2B5B02002478).There was no additional external funding received for this study.
文摘Early detection of the Covid-19 disease is essential due to its higher rate of infection affecting tens of millions of people,and its high number of deaths also by 7%.For that purpose,a proposed model of several stages was developed.The first stage is optimizing the images using dynamic adaptive histogram equalization,performing a semantic segmentation using DeepLabv3Plus,then augmenting the data by flipping it horizontally,rotating it,then flipping it vertically.The second stage builds a custom convolutional neural network model using several pre-trained ImageNet.Finally,the model compares the pre-trained data to the new output,while repeatedly trimming the best-performing models to reduce complexity and improve memory efficiency.Several experiments were done using different techniques and parameters.Accordingly,the proposed model achieved an average accuracy of 99.6%and an area under the curve of 0.996 in the Covid-19 detection.This paper will discuss how to train a customized intelligent convolutional neural network using various parameters on a set of chest X-rays with an accuracy of 99.6%.
文摘The Wavelet-Domain Projection Pursuit Learning Network (WDPPLN) is proposedfor restoring degraded image. The new network combines the advantages of both projectionpursuit and wavelet shrinkage. Restoring image is very difficult when little is known about apriori knowledge for multisource degraded factors. WDPPLN successfully resolves this problemby separately processing wavelet coefficients and scale coefficients. Parameters in WDPPLN,which are used to simulate degraded factors, are estimated via WDPPLN training, using scalecoefficients. Also, WDPPLN uses soft-threshold of wavelet shrinkage technique to suppress noisein three high frequency subbands. The new method is compared with the traditional methodsand the Projection Pursuit Learning Network (PPLN) method. Experimental results demonstratethat it is an effective method for unsupervised restoring degraded image.
基金supported by a State Grid Zhejiang Electric Power Co.,Ltd.Economic and Technical Research Institute Project(Key Technologies and Empirical Research of Diversified Integrated Operation of User-Side Energy Storage in Power Market Environment,No.5211JY19000W)supported by the National Natural Science Foundation of China(Research on Power Market Management to Promote Large-Scale New Energy Consumption,No.71804045).
文摘In the electricity market,fluctuations in real-time prices are unstable,and changes in short-term load are determined by many factors.By studying the timing of charging and discharging,as well as the economic benefits of energy storage in the process of participating in the power market,this paper takes energy storage scheduling as merely one factor affecting short-term power load,which affects short-term load time series along with time-of-use price,holidays,and temperature.A deep learning network is used to predict the short-term load,a convolutional neural network(CNN)is used to extract the features,and a long short-term memory(LSTM)network is used to learn the temporal characteristics of the load value,which can effectively improve prediction accuracy.Taking the load data of a certain region as an example,the CNN-LSTM prediction model is compared with the single LSTM prediction model.The experimental results show that the CNN-LSTM deep learning network with the participation of energy storage in dispatching can have high prediction accuracy for short-term power load forecasting.
基金Projects(42174170,41874145,72088101)supported by the National Natural Science Foundation of ChinaProject(CX20200228)supported by the Hunan Provincial Innovation Foundation for Postgraduate,China。
文摘Landfill leaks pose a serious threat to environmental health,risking the contamination of both groundwater and soil resources.Accurate investigation of these sites is essential for implementing effective prevention and control measures.The self-potential(SP)stands out for its sensitivity to contamination plumes,offering a solution for monitoring and detecting the movement and seepage of subsurface pollutants.However,traditional SP inversion techniques heavily rely on precise subsurface resistivity information.In this study,we propose the Attention U-Net deep learning network for rapid SP inversion.By incorporating an attention mechanism,this algorithm effectively learns the relationship between array-style SP data and the location and extent of subsurface contaminated sources.We designed a synthetic landfill model with a heterogeneous resistivity structure to assess the performance of Attention U-Net deep learning network.Additionally,we conducted further validation using a laboratory model to assess its practical applicability.The results demonstrate that the algorithm is not solely dependent on resistivity information,enabling effective locating of the source distribution,even in models with intricate subsurface structures.Our work provides a promising tool for SP data processing,enhancing the applicability of this method in the field of near-subsurface environmental monitoring.
文摘This study investigated how students used peer assessments in synchronous learning network (SLN) to assess each other s writing. It focused on examining the frequency and styles of various techniques students employed while assessing each others writing and student response to assessing each other s writing in a SLN context. The findings indicated that these students received many assessments during each peer assessment activity. They preferred to use assessing techniques of less critical types, and had po...
文摘The transparent open box(TOB)learning network algorithm offers an alternative approach to the lack of transparency provided by most machine-learning algorithms.It provides the exact calculations and relationships among the underlying input variables of the datasets to which it is applied.It also has the capability to achieve credible and auditable levels of prediction accuracy to complex,non-linear datasets,typical of those encountered in the oil and gas sector,highlighting the potential for underfitting and overfitting.The algorithm is applied here to predict bubble-point pressure from a published PVT dataset of 166 data records involving four easy-tomeasure variables(reservoir temperature,gas-oil ratio,oil gravity,gas density relative to air)with uneven,and in parts,sparse data coverage.The TOB network demonstrates high-prediction accuracy for this complex system,although it predictions applied to the full dataset are outperformed by an artificial neural network(ANN).However,the performance of the TOB algorithm reveals the risk of overfitting in the sparse areas of the dataset and achieves a prediction performance that matches the ANN algorithm where the underlying data population is adequate.The high levels of transparency and its inhibitions to overfitting enable the TOB learning network to provide complementary information about the underlying dataset to that provided by traditional machine learning algorithms.This makes them suitable for application in parallel with neural-network algorithms,to overcome their black-box tendencies,and for benchmarking the prediction performance of other machine learning algorithms.
文摘Purpose-Parkinson’s disease(PD)is a well-known complex neurodegenerative disease.Typically,its identification is based on motor disorders,while the computer estimation of its main symptoms with computational machine learning(ML)has a high exposure which is supported by researches conducted.Nevertheless,ML approaches required first to refine their parameters and then to work with the best model generated.This process often requires an expert user to oversee the performance of the algorithm.Therefore,an attention is required towards new approaches for better forecasting accuracy.Design/methodology/approach-To provide an available identification model for Parkinson disease as an auxiliary function for clinicians,the authors suggest a new evolutionary classification model.The core of the prediction model is a fast learning network(FLN)optimized by a genetic algorithm(GA).To get a better subset of features and parameters,a new coding architecture is introduced to improve GA for obtaining an optimal FLN model.Findings-The proposed model is intensively evaluated through a series of experiments based on Speech and HandPD benchmark datasets.The very popular wrappers induction models such as support vector machine(SVM),K-nearest neighbors(KNN)have been tested in the same condition.The results support that the proposed model can achieve the best performances in terms of accuracy and g-mean.Originality/value-A novel efficient PD detectionmodel is proposed,which is called A-W-FLN.The A-W-FLN utilizes FLN as the base classifier;in order to take its higher generalization ability,and identification capability is alsoembedded to discover themost suitable featuremodel in the detection process.Moreover,the proposedmethod automatically optimizes the FLN’s architecture to a smaller number of hidden nodes and solid connecting weights.This helps the network to train on complex PD datasets with non-linear features and yields superior result.
基金This work was partially funded by the Industry Grant Scheme from Jaycorp Berhad in cooperation with UNITAR International University.The authors would like to thank INSFORNET,the Center for Advanced Computing Technology(C-ACT)at Universiti Teknikal Malaysia Melaka(UTeM),and the Center of Intelligent and Autonomous Systems(CIAS)at Universiti Tun Hussein Onn Malaysia(UTHM)for supporting this work.
文摘With an increasing number of services connected to the internet,including cloud computing and Internet of Things(IoT)systems,the prevention of cyberattacks has become more challenging due to the high dimensionality of the network traffic data and access points.Recently,researchers have suggested deep learning(DL)algorithms to define intrusion features through training empirical data and learning anomaly patterns of attacks.However,due to the high dynamics and imbalanced nature of the data,the existing DL classifiers are not completely effective at distinguishing between abnormal and normal behavior line connections for modern networks.Therefore,it is important to design a self-adaptive model for an intrusion detection system(IDS)to improve the detection of attacks.Consequently,in this paper,a novel hybrid weighted deep belief network(HW-DBN)algorithm is proposed for building an efficient and reliable IDS(DeepIoT.IDS)model to detect existing and novel cyberattacks.The HW-DBN algorithm integrates an improved Gaussian–Bernoulli restricted Boltzmann machine(Deep GB-RBM)feature learning operator with a weighted deep neural networks(WDNN)classifier.The CICIDS2017 dataset is selected to evaluate the DeepIoT.IDS model as it contains multiple types of attacks,complex data patterns,noise values,and imbalanced classes.We have compared the performance of the DeepIoT.IDS model with three recent models.The results show the DeepIoT.IDS model outperforms the three other models by achieving a higher detection accuracy of 99.38%and 99.99%for web attack and bot attack scenarios,respectively.Furthermore,it can detect the occurrence of low-frequency attacks that are undetectable by other models.
基金supported by the National Natural Science Foundation of China (60974082,11171094)the Fundamental Research Funds for the Central Universities (K50510700004)+1 种基金the Foundation and Advanced Technology Research Program of Henan Province (102300410264)the Basic Research Program of the Education Department of Henan Province (2010A110010)
文摘Structure learning of Bayesian networks is a wellresearched but computationally hard task.For learning Bayesian networks,this paper proposes an improved algorithm based on unconstrained optimization and ant colony optimization(U-ACO-B) to solve the drawbacks of the ant colony optimization(ACO-B).In this algorithm,firstly,an unconstrained optimization problem is solved to obtain an undirected skeleton,and then the ACO algorithm is used to orientate the edges,thus returning the final structure.In the experimental part of the paper,we compare the performance of the proposed algorithm with ACO-B algorithm.The experimental results show that our method is effective and greatly enhance convergence speed than ACO-B algorithm.
文摘To get simpler operation in modified fuzzy adaptive learning control network (FALCON) in some engineering application, sigmoid nonlinear function is employed as a substitute of traditional Gaussian membership function. For making the modified FALCON learning more efficient and stable, a simulated annealing (SA) learning coefficient is introduced into learning algorithm. At first, the basic concepts and main advantages of FALCON were briefly reviewed. Subsequently, the topological structure and nodes operation were illustrated; the gradient-descent learning algorithm with SA learning coefficient was derived; and the distinctions between the archetype and the modification were analyzed. Eventually, the significance and worthiness of the modified FALCON were validated by its application to probability prediction of anode effect in aluminium electrolysis cells.
基金supported by the Natural Science Foundation of Tianjin(No.20JCQNJC00720)the Fundamental Research Fund for the Central Universities(No.3122021052)。
文摘The homogeneity analysis of multi-airport system can provide important decision-making support for the route layout and cooperative operation.Existing research seldom analyzes the homogeneity of multi-airport system from the perspective of route network analysis,and the attribute information of airport nodes in the airport route network is not appropriately integrated into the airport network.In order to solve this problem,a multi-airport system homogeneity analysis method based on airport attribute network representation learning is proposed.Firstly,the route network of a multi-airport system with attribute information is constructed.If there are flights between airports,an edge is added between airports,and regional attribute information is added for each airport node.Secondly,the airport attributes and the airport network vector are represented respectively.The airport attributes and the airport network vector are embedded into the unified airport representation vector space by the network representation learning method,and then the airport vector integrating the airport attributes and the airport network characteristics is obtained.By calculating the similarity of the airport vectors,it is convenient to calculate the degree of homogeneity between airports and the homogeneity of the multi-airport system.The experimental results on the Beijing-Tianjin-Hebei multi-airport system show that,compared with other existing algorithms,the homogeneity analysis method based on attributed network representation learning can get more consistent results with the current situation of Beijing-Tianjin-Hebei multi-airport system.
基金supported by National Natural Science Foundation of China (No.60970055)
文摘Frequent counting is a very so often required operation in machine learning algorithms. A typical machine learning task, learning the structure of Bayesian network (BN) based on metric scoring, is introduced as an example that heavily relies on frequent counting. A fast calculation method for frequent counting enhanced with two cache layers is then presented for learning BN. The main contribution of our approach is to eliminate comparison operations for frequent counting by introducing a multi-radix number system calculation. Both mathematical analysis and empirical comparison between our method and state-of-the-art solution are conducted. The results show that our method is dominantly superior to state-of-the-art solution in solving the problem of learning BN.
基金the National Research Foundation of Korea(NRF)grant of the Korea government(MSIP)(2020R1A2B5B01001899)(Grantee:GJY,http://www.nrf.re.kr)and Institute of Engineering Research at Seoul National University(Grantee:GJY,http://www.snu.ac.kr).The authors are grateful for their supports.
文摘This paper first proposes a new self-learning data-driven methodology that can develop the failure criteria of unknown anisotropic ductile materials from the minimal number of experimental tests.Establishing failure criteria of anisotropic ductile materials requires time-consuming tests and manual data evaluation.The proposed method can overcome such practical challenges.The methodology is formalized by combining four ideas:1)The deep learning neural network(DLNN)-based material constitutive model,2)Self-learning inverse finite element(SELIFE)simulation,3)Algorithmic identification of failure points from the selflearned stress-strain curves and 4)Derivation of the failure criteria through symbolic regression of the genetic programming.Stress update and the algorithmic tangent operator were formulated in terms of DLNN parameters for nonlinear finite element analysis.Then,the SELIFE simulation algorithm gradually makes the DLNN model learn highly complex multi-axial stress and strain relationships,being guided by the experimental boundary measurements.Following the failure point identification,a self-learning data-driven failure criteria are eventually developed with the help of a reliable symbolic regression algorithm.The methodology and the self-learning data-driven failure criteria were verified by comparing with a reference failure criteria and simulating with different materials orientations,respectively.
基金What is more,we thank the National Natural Science Foundation of China(Nos.61966039,62241604)the Scientific Research Fund Project of the Education Department of Yunnan Province(No.2023Y0565)Also,this work was supported in part by the Xingdian Talent Support Program for Young Talents(No.XDYC-QNRC-2022-0518).
文摘Many network presentation learning algorithms(NPLA)have originated from the process of the random walk between nodes in recent years.Despite these algorithms can obtain great embedding results,there may be also some limitations.For instance,only the structural information of nodes is considered when these kinds of algorithms are constructed.Aiming at this issue,a label and community information-based network presentation learning algorithm(LC-NPLA)is proposed in this paper.First of all,by using the community information and the label information of nodes,the first-order neighbors of nodes are reconstructed.In the next,the random walk strategy is improved by integrating the degree information and label information of nodes.Then,the node sequence obtained from random walk sampling is transformed into the node representation vector by the Skip-Gram model.At last,the experimental results on ten real-world networks demonstrate that the proposed algorithm has great advantages in the label classification,network reconstruction and link prediction tasks,compared with three benchmark algorithms.
基金supported by CITRIS and the Banatao Institute,Air Force Office of Scientific Research(Grant No.FA9550-22-1-0420)National Science Foundation(Grant No.ACI-1548562).
文摘Conventional wing aerodynamic optimization processes can be time-consuming and imprecise due to the complexity of versatile flight missions.Plenty of existing literature has considered two-dimensional infinite airfoil optimization,while three-dimensional finite wing optimizations are subject to limited study because of high computational costs.Here we create an adaptive optimization methodology built upon digitized wing shape deformation and deep learning algorithms,which enable the rapid formulation of finite wing designs for specific aerodynamic performance demands under different cruise conditions.This methodology unfolds in three stages:radial basis function interpolated wing generation,collection of inputs from computational fluid dynamics simulations,and deep neural network that constructs the surrogate model for the optimal wing configuration.It has been demonstrated that the proposed methodology can significantly reduce the computational cost of numerical simulations.It also has the potential to optimize various aerial vehicles undergoing different mission environments,loading conditions,and safety requirements.
基金supported by the National Science and Technology Major Project of the Ministry of Science and Technology of China(2014 ZX03001027)
文摘There are various heterogeneous networks for terminals to deliver a better quality of service. Signal system recognition and classification contribute a lot to the process. However, in low signal to noise ratio(SNR) circumstances or under time-varying multipath channels, the majority of the existing algorithms for signal recognition are already facing limitations. In this series, we present a robust signal recognition method based upon the original and latest updated version of the extreme learning machine(ELM) to help users to switch between networks. The ELM utilizes signal characteristics to distinguish systems. The superiority of this algorithm lies in the random choices of hidden nodes and in the fact that it determines the output weights analytically, which result in lower complexity. Theoretically, the algorithm tends to offer a good generalization performance at an extremely fast speed of learning. Moreover, we implement the GSM/WCDMA/LTE models in the Matlab environment by using the Simulink tools. The simulations reveal that the signals can be recognized successfully to achieve a 95% accuracy in a low SNR(0 dB) environment in the time-varying multipath Rayleigh fading channel.
基金This work was supported by the Kyonggi University Research Grant 2022.
文摘Recommendation Information Systems(RIS)are pivotal in helping users in swiftly locating desired content from the vast amount of information available on the Internet.Graph Convolution Network(GCN)algorithms have been employed to implement the RIS efficiently.However,the GCN algorithm faces limitations in terms of performance enhancement owing to the due to the embedding value-vanishing problem that occurs during the learning process.To address this issue,we propose a Weighted Forwarding method using the GCN(WF-GCN)algorithm.The proposed method involves multiplying the embedding results with different weights for each hop layer during graph learning.By applying the WF-GCN algorithm,which adjusts weights for each hop layer before forwarding to the next,nodes with many neighbors achieve higher embedding values.This approach facilitates the learning of more hop layers within the GCN framework.The efficacy of the WF-GCN was demonstrated through its application to various datasets.In the MovieLens dataset,the implementation of WF-GCN in LightGCN resulted in significant performance improvements,with recall and NDCG increasing by up to+163.64%and+132.04%,respectively.Similarly,in the Last.FM dataset,LightGCN using WF-GCN enhanced with WF-GCN showed substantial improvements,with the recall and NDCG metrics rising by up to+174.40%and+169.95%,respectively.Furthermore,the application of WF-GCN to Self-supervised Graph Learning(SGL)and Simple Graph Contrastive Learning(SimGCL)also demonstrated notable enhancements in both recall and NDCG across these datasets.
基金supported by two Ministry of Education(MoE)Singapore Tier 1 research grants under grant numbers R-296-000-208-133 and R-296-000-241-114.
文摘Computer-empowered detection of possible faults for Heating,Ventilation and Air-Conditioning(HVAC)subsystems,e.g.,chillers,is one of the most important applications in Artificial Intelligence(AI)integrated Internet of Things(IoT).The cyber-physical system greatly enhances the safety and security of the working facilities,reducing time,saving energy and protecting humans’health.Under the current trends of smart building design and energy management optimization,Automated Fault Detection and Diagnosis(AFDD)of chillers integrated with IoT is highly demanded.Recent studies show that standard machine learning techniques,such as Principal Component Analysis(PCA),Support Vector Machine(SVM)and tree-structure-based algorithms,are useful in capturing various chiller faults with high accuracy rates.With the fast development of deep learning technology,Convolutional Neural Networks(CNNs)have been widely and successfully applied to various fields.However,for chiller AFDD,few existing works are adopting CNN and its extensions in the feature extraction and classification processes.In this study,we propose to perform chiller FDD using a CNN-based approach.The proposed approach has two distinct advantages over existing machine learning-based chiller AFDD methods.First,the CNN-based approach does not require the feature selection/extraction process.Since CNN is reputable with its feature extraction capability,the feature extraction and classification processes are merged,leading to a more neat AFDD framework compared to traditional approaches.Second,the classification accuracy is significantly improved compared to traditional methods using the CNN-based approach.
基金funded by Vietnam National Foundation for Science and Technology Development (NAFOSTED)under Grant No.105.08-2019.03.
文摘Flash floods are one of the most dangerous natural disasters,especially in hilly terrain,causing loss of life,property,and infrastructures and sudden disruption of traffic.These types of floods are mostly associated with landslides and erosion of roads within a short time.Most of Vietnamis hilly and mountainous;thus,the problem due to flash flood is severe and requires systematic studies to correctly identify flood susceptible areas for proper landuse planning and traffic management.In this study,three Machine Learning(ML)methods namely Deep Learning Neural Network(DL),Correlation-based FeatureWeighted Naive Bayes(CFWNB),and Adaboost(AB-CFWNB)were used for the development of flash flood susceptibility maps for hilly road section(115 km length)of National Highway(NH)-6 inHoa Binh province,Vietnam.In the proposedmodels,88 past flash flood events were used together with 14 flash floods affecting topographical and geo-environmental factors.The performance of themodels was evaluated using standard statisticalmeasures including Receiver Operating Characteristic(ROC)Curve,Area Under Curve(AUC)and Root Mean Square Error(RMSE).The results revealed that all the models performed well(AUC>0.80)in predicting flash flood susceptibility zones,but the performance of the DL model is the best(AUC:0.972,RMSE:0.352).Therefore,the DL model can be applied to develop an accurate flash flood susceptibility map of hilly terrain which can be used for proper planning and designing of the highways and other infrastructure facilities besides landuse management of the area.
基金supported in part by the National Natural Science Foundation of China(62022014)in part by the National Key Research and Development Program of China(2017YFE0117000)。
文摘Existing biomimetic robots can perform some basic rat-like movement primitives(MPs)and simple behavior with stiff combinations of these MPs.To mimic typical rat behavior with high similarity,we propose parameterizing the behavior using a probabilistic model and movement characteristics.First,an analysis of fifteen 10 min video sequences revealed that an actual rat has six typical behaviors in the open field,and each kind of behavior contains different bio-inspired combinations of eight MPs.We used the softmax classifier to obtain the behavior-movement hierarchical probability model.Secondly,we specified the MPs using movement parameters that are static and dynamic.We obtained the predominant values of the static and dynamic movement parameters using hierarchical clustering and fuzzy C-means clustering,respectively.These predominant parameters were used for fitting the rat spinal joint trajectory using a second-order Fourier series,and the joint trajectory was generalized using a back propagation neural network with two hidden layers.Finally,the hierarchical probability model and the generalized joint trajectory were mapped to the robot as control policy and commands,respectively.We implemented the six typical behaviors on the robot,and the results show high similarity when compared with the behaviors of actual rats.