期刊文献+
共找到265,592篇文章
< 1 2 250 >
每页显示 20 50 100
Rao Algorithms-Based Structure Optimization for Heterogeneous Wireless Sensor Networks 被引量:1
1
作者 Shereen K.Refaay Samia A.Ali +2 位作者 Moumen T.El-Melegy Louai A.Maghrabi Hamdy H.El-Sayed 《Computers, Materials & Continua》 SCIE EI 2024年第1期873-897,共25页
The structural optimization of wireless sensor networks is a critical issue because it impacts energy consumption and hence the network’s lifetime.Many studies have been conducted for homogeneous networks,but few hav... The structural optimization of wireless sensor networks is a critical issue because it impacts energy consumption and hence the network’s lifetime.Many studies have been conducted for homogeneous networks,but few have been performed for heterogeneouswireless sensor networks.This paper utilizes Rao algorithms to optimize the structure of heterogeneous wireless sensor networks according to node locations and their initial energies.The proposed algorithms lack algorithm-specific parameters and metaphorical connotations.The proposed algorithms examine the search space based on the relations of the population with the best,worst,and randomly assigned solutions.The proposed algorithms can be evaluated using any routing protocol,however,we have chosen the well-known routing protocols in the literature:Low Energy Adaptive Clustering Hierarchy(LEACH),Power-Efficient Gathering in Sensor Information Systems(PEAGSIS),Partitioned-based Energy-efficient LEACH(PE-LEACH),and the Power-Efficient Gathering in Sensor Information Systems Neural Network(PEAGSIS-NN)recent routing protocol.We compare our optimized method with the Jaya,the Particle Swarm Optimization-based Energy Efficient Clustering(PSO-EEC)protocol,and the hybrid Harmony Search Algorithm and PSO(HSA-PSO)algorithms.The efficiencies of our proposed algorithms are evaluated by conducting experiments in terms of the network lifetime(first dead node,half dead nodes,and last dead node),energy consumption,packets to cluster head,and packets to the base station.The experimental results were compared with those obtained using the Jaya optimization algorithm.The proposed algorithms exhibited the best performance.The proposed approach successfully prolongs the network lifetime by 71% for the PEAGSIS protocol,51% for the LEACH protocol,10% for the PE-LEACH protocol,and 73% for the PEGSIS-NN protocol;Moreover,it enhances other criteria such as energy conservation,fitness convergence,packets to cluster head,and packets to the base station. 展开更多
关键词 Wireless sensor networks Rao algorithms OPTIMIZATION LEACH PEAGSIS
下载PDF
Optimizing Connections:Applied Shortest Path Algorithms for MANETs
2
作者 Ibrahim Alameri Jitka Komarkova +2 位作者 Tawfik Al-Hadhrami Abdulsamad Ebrahim Yahya Atef Gharbi 《Computer Modeling in Engineering & Sciences》 SCIE EI 2024年第10期787-807,共21页
This study is trying to address the critical need for efficient routing in Mobile Ad Hoc Networks(MANETs)from dynamic topologies that pose great challenges because of the mobility of nodes.Themain objective was to del... This study is trying to address the critical need for efficient routing in Mobile Ad Hoc Networks(MANETs)from dynamic topologies that pose great challenges because of the mobility of nodes.Themain objective was to delve into and refine the application of the Dijkstra’s algorithm in this context,a method conventionally esteemed for its efficiency in static networks.Thus,this paper has carried out a comparative theoretical analysis with the Bellman-Ford algorithm,considering adaptation to the dynamic network conditions that are typical for MANETs.This paper has shown through detailed algorithmic analysis that Dijkstra’s algorithm,when adapted for dynamic updates,yields a very workable solution to the problem of real-time routing in MANETs.The results indicate that with these changes,Dijkstra’s algorithm performs much better computationally and 30%better in routing optimization than Bellman-Ford when working with configurations of sparse networks.The theoretical framework adapted,with the adaptation of the Dijkstra’s algorithm for dynamically changing network topologies,is novel in this work and quite different from any traditional application.The adaptation should offer more efficient routing and less computational overhead,most apt in the limited resource environment of MANETs.Thus,from these findings,one may derive a conclusion that the proposed version of Dijkstra’s algorithm is the best and most feasible choice of the routing protocol for MANETs given all pertinent key performance and resource consumption indicators and further that the proposed method offers a marked improvement over traditional methods.This paper,therefore,operationalizes the theoretical model into practical scenarios and also further research with empirical simulations to understand more about its operational effectiveness. 展开更多
关键词 Dijkstra’s algorithm optimization complexity analysis shortest path first comparative algorithm analysis nondeterministic polynomial(NP)-complete
下载PDF
Intelligent Solution System for Cloud Security Based on Equity Distribution:Model and Algorithms
3
作者 Sarah Mustafa Eljack Mahdi Jemmali +3 位作者 Mohsen Denden Mutasim Al Sadig Abdullah M.Algashami Sadok Turki 《Computers, Materials & Continua》 SCIE EI 2024年第1期1461-1479,共19页
In the cloud environment,ensuring a high level of data security is in high demand.Data planning storage optimization is part of the whole security process in the cloud environment.It enables data security by avoiding ... In the cloud environment,ensuring a high level of data security is in high demand.Data planning storage optimization is part of the whole security process in the cloud environment.It enables data security by avoiding the risk of data loss and data overlapping.The development of data flow scheduling approaches in the cloud environment taking security parameters into account is insufficient.In our work,we propose a data scheduling model for the cloud environment.Themodel is made up of three parts that together help dispatch user data flow to the appropriate cloudVMs.The first component is the Collector Agent whichmust periodically collect information on the state of the network links.The second one is the monitoring agent which must then analyze,classify,and make a decision on the state of the link and finally transmit this information to the scheduler.The third one is the scheduler who must consider previous information to transfer user data,including fair distribution and reliable paths.It should be noted that each part of the proposedmodel requires the development of its algorithms.In this article,we are interested in the development of data transfer algorithms,including fairness distribution with the consideration of a stable link state.These algorithms are based on the grouping of transmitted files and the iterative method.The proposed algorithms showthe performances to obtain an approximate solution to the studied problem which is an NP-hard(Non-Polynomial solution)problem.The experimental results show that the best algorithm is the half-grouped minimum excluding(HME),with a percentage of 91.3%,an average deviation of 0.042,and an execution time of 0.001 s. 展开更多
关键词 Cyber-security cloud computing cloud security algorithms HEURISTICS
下载PDF
Topological optimization of ballistic protective structures through genetic algorithms in a vulnerability-driven environment
4
作者 Salvatore Annunziata Luca Lomazzi +1 位作者 Marco Giglio Andrea Manes 《Defence Technology(防务技术)》 SCIE EI CAS CSCD 2024年第10期125-137,共13页
Reducing the vulnerability of a platform,i.e.,the risk of being affected by hostile objects,is of paramount importance in the design process of vehicles,especially aircraft.A simple and effective way to decrease vulne... Reducing the vulnerability of a platform,i.e.,the risk of being affected by hostile objects,is of paramount importance in the design process of vehicles,especially aircraft.A simple and effective way to decrease vulnerability is to introduce protective structures to intercept and possibly stop threats.However,this type of solution can lead to a significant increase in weight,affecting the performance of the aircraft.For this reason,it is crucial to study possible solutions that allow reducing the vulnerability of the aircraft while containing the increase in structural weight.One possible strategy is to optimize the topology of protective solutions to find the optimal balance between vulnerability and the weight of the added structures.Among the many optimization techniques available in the literature for this purpose,multiobjective genetic algorithms stand out as promising tools.In this context,this work proposes the use of a in-house software for vulnerability calculation to guide the process of topology optimization through multi-objective genetic algorithms,aiming to simultaneously minimize the weight of protective structures and vulnerability.In addition to the use of the in-house software,which itself represents a novelty in the field of topology optimization of structures,the method incorporates a custom mutation function within the genetic algorithm,specifically developed using a graph-based approach to ensure the continuity of the generated structures.The tool developed for this work is capable of generating protections with optimized layouts considering two different types of impacting objects,namely bullets and fragments from detonating objects.The software outputs a set of non-dominated solutions describing different topologies that the user can choose from. 展开更多
关键词 Topological optimization Protective structure Genetic algorithm SURVIVABILITY VULNERABILITY
下载PDF
An Online Fake Review Detection Approach Using Famous Machine Learning Algorithms
5
作者 Asma Hassan Alshehri 《Computers, Materials & Continua》 SCIE EI 2024年第2期2767-2786,共20页
Online review platforms are becoming increasingly popular,encouraging dishonest merchants and service providers to deceive customers by creating fake reviews for their goods or services.Using Sybil accounts,bot farms,... Online review platforms are becoming increasingly popular,encouraging dishonest merchants and service providers to deceive customers by creating fake reviews for their goods or services.Using Sybil accounts,bot farms,and real account purchases,immoral actors demonize rivals and advertise their goods.Most academic and industry efforts have been aimed at detecting fake/fraudulent product or service evaluations for years.The primary hurdle to identifying fraudulent reviews is the lack of a reliable means to distinguish fraudulent reviews from real ones.This paper adopts a semi-supervised machine learning method to detect fake reviews on any website,among other things.Online reviews are classified using a semi-supervised approach(PU-learning)since there is a shortage of labeled data,and they are dynamic.Then,classification is performed using the machine learning techniques Support Vector Machine(SVM)and Nave Bayes.The performance of the suggested system has been compared with standard works,and experimental findings are assessed using several assessment metrics. 展开更多
关键词 SECURITY fake review semi-supervised learning ML algorithms review detection
下载PDF
Uniaxial Compressive Strength Prediction for Rock Material in Deep Mine Using Boosting-Based Machine Learning Methods and Optimization Algorithms
6
作者 Junjie Zhao Diyuan Li +1 位作者 Jingtai Jiang Pingkuang Luo 《Computer Modeling in Engineering & Sciences》 SCIE EI 2024年第7期275-304,共30页
Traditional laboratory tests for measuring rock uniaxial compressive strength(UCS)are tedious and timeconsuming.There is a pressing need for more effective methods to determine rock UCS,especially in deep mining envir... Traditional laboratory tests for measuring rock uniaxial compressive strength(UCS)are tedious and timeconsuming.There is a pressing need for more effective methods to determine rock UCS,especially in deep mining environments under high in-situ stress.Thus,this study aims to develop an advanced model for predicting the UCS of rockmaterial in deepmining environments by combining three boosting-basedmachine learning methods with four optimization algorithms.For this purpose,the Lead-Zinc mine in Southwest China is considered as the case study.Rock density,P-wave velocity,and point load strength index are used as input variables,and UCS is regarded as the output.Subsequently,twelve hybrid predictive models are obtained.Root mean square error(RMSE),mean absolute error(MAE),coefficient of determination(R2),and the proportion of the mean absolute percentage error less than 20%(A-20)are selected as the evaluation metrics.Experimental results showed that the hybridmodel consisting of the extreme gradient boostingmethod and the artificial bee colony algorithm(XGBoost-ABC)achieved satisfactory results on the training dataset and exhibited the best generalization performance on the testing dataset.The values of R2,A-20,RMSE,and MAE on the training dataset are 0.98,1.0,3.11 MPa,and 2.23MPa,respectively.The highest values of R2 and A-20(0.93 and 0.96),and the smallest RMSE and MAE values of 4.78 MPa and 3.76MPa,are observed on the testing dataset.The proposed hybrid model can be considered a reliable and effective method for predicting rock UCS in deep mines. 展开更多
关键词 Uniaxial compression strength strength prediction machine learning optimization algorithm
下载PDF
Data-Driven Learning Control Algorithms for Unachievable Tracking Problems
7
作者 Zeyi Zhang Hao Jiang +1 位作者 Dong Shen Samer S.Saab 《IEEE/CAA Journal of Automatica Sinica》 SCIE EI CSCD 2024年第1期205-218,共14页
For unachievable tracking problems, where the system output cannot precisely track a given reference, achieving the best possible approximation for the reference trajectory becomes the objective. This study aims to in... For unachievable tracking problems, where the system output cannot precisely track a given reference, achieving the best possible approximation for the reference trajectory becomes the objective. This study aims to investigate solutions using the Ptype learning control scheme. Initially, we demonstrate the necessity of gradient information for achieving the best approximation.Subsequently, we propose an input-output-driven learning gain design to handle the imprecise gradients of a class of uncertain systems. However, it is discovered that the desired performance may not be attainable when faced with incomplete information.To address this issue, an extended iterative learning control scheme is introduced. In this scheme, the tracking errors are modified through output data sampling, which incorporates lowmemory footprints and offers flexibility in learning gain design.The input sequence is shown to converge towards the desired input, resulting in an output that is closest to the given reference in the least square sense. Numerical simulations are provided to validate the theoretical findings. 展开更多
关键词 Data-driven algorithms incomplete information iterative learning control gradient information unachievable problems
下载PDF
Landslide susceptibility mapping(LSM)based on different boosting and hyperparameter optimization algorithms:A case of Wanzhou District,China
8
作者 Deliang Sun Jing Wang +2 位作者 Haijia Wen YueKai Ding Changlin Mi 《Journal of Rock Mechanics and Geotechnical Engineering》 SCIE CSCD 2024年第8期3221-3232,共12页
Boosting algorithms have been widely utilized in the development of landslide susceptibility mapping(LSM)studies.However,these algorithms possess distinct computational strategies and hyperparameters,making it challen... Boosting algorithms have been widely utilized in the development of landslide susceptibility mapping(LSM)studies.However,these algorithms possess distinct computational strategies and hyperparameters,making it challenging to propose an ideal LSM model.To investigate the impact of different boosting algorithms and hyperparameter optimization algorithms on LSM,this study constructed a geospatial database comprising 12 conditioning factors,such as elevation,stratum,and annual average rainfall.The XGBoost(XGB),LightGBM(LGBM),and CatBoost(CB)algorithms were employed to construct the LSM model.Furthermore,the Bayesian optimization(BO),particle swarm optimization(PSO),and Hyperband optimization(HO)algorithms were applied to optimizing the LSM model.The boosting algorithms exhibited varying performances,with CB demonstrating the highest precision,followed by LGBM,and XGB showing poorer precision.Additionally,the hyperparameter optimization algorithms displayed different performances,with HO outperforming PSO and BO showing poorer performance.The HO-CB model achieved the highest precision,boasting an accuracy of 0.764,an F1-score of 0.777,an area under the curve(AUC)value of 0.837 for the training set,and an AUC value of 0.863 for the test set.The model was interpreted using SHapley Additive exPlanations(SHAP),revealing that slope,curvature,topographic wetness index(TWI),degree of relief,and elevation significantly influenced landslides in the study area.This study offers a scientific reference for LSM and disaster prevention research.This study examines the utilization of various boosting algorithms and hyperparameter optimization algorithms in Wanzhou District.It proposes the HO-CB-SHAP framework as an effective approach to accurately forecast landslide disasters and interpret LSM models.However,limitations exist concerning the generalizability of the model and the data processing,which require further exploration in subsequent studies. 展开更多
关键词 Landslide susceptibility Hyperparameter optimization Boosting algorithms SHapley additive exPlanations(SHAP)
下载PDF
Structure-preserving algorithms for guiding center dynamics based on the slow manifold of classical Pauli particle
9
作者 张若涵 王正汹 +1 位作者 肖建元 王丰 《Plasma Science and Technology》 SCIE EI CAS CSCD 2024年第6期88-102,共15页
The classical Pauli particle(CPP) serves as a slow manifold, substituting the conventional guiding center dynamics. Based on the CPP, we utilize the averaged vector field(AVF) method in the computations of drift orbit... The classical Pauli particle(CPP) serves as a slow manifold, substituting the conventional guiding center dynamics. Based on the CPP, we utilize the averaged vector field(AVF) method in the computations of drift orbits. Demonstrating significantly higher efficiency, this advanced method is capable of accomplishing the simulation in less than one-third of the time of directly computing the guiding center motion. In contrast to the CPP-based Boris algorithm, this approach inherits the advantages of the AVF method, yielding stable trajectories even achieved with a tenfold time step and reducing the energy error by two orders of magnitude. By comparing these two CPP algorithms with the traditional RK4 method, the numerical results indicate a remarkable performance in terms of both the computational efficiency and error elimination. Moreover, we verify the properties of slow manifold integrators and successfully observe the bounce on both sides of the limiting slow manifold with deliberately chosen perturbed initial conditions. To evaluate the practical value of the methods, we conduct simulations in non-axisymmetric perturbation magnetic fields as part of the experiments,demonstrating that our CPP-based AVF method can handle simulations under complex magnetic field configurations with high accuracy, which the CPP-based Boris algorithm lacks. Through numerical experiments, we demonstrate that the CPP can replace guiding center dynamics in using energy-preserving algorithms for computations, providing a new, efficient, as well as stable approach for applying structure-preserving algorithms in plasma simulations. 展开更多
关键词 structure-preserving algorithm averaged vector field classical Pauli particle guiding center dynamics
下载PDF
A Comparative Study of Metaheuristic Optimization Algorithms for Solving Real-World Engineering Design Problems
10
作者 Elif Varol Altay Osman Altay Yusuf Ovik 《Computer Modeling in Engineering & Sciences》 SCIE EI 2024年第4期1039-1094,共56页
Real-world engineering design problems with complex objective functions under some constraints are relatively difficult problems to solve.Such design problems are widely experienced in many engineering fields,such as ... Real-world engineering design problems with complex objective functions under some constraints are relatively difficult problems to solve.Such design problems are widely experienced in many engineering fields,such as industry,automotive,construction,machinery,and interdisciplinary research.However,there are established optimization techniques that have shown effectiveness in addressing these types of issues.This research paper gives a comparative study of the implementation of seventeen new metaheuristic methods in order to optimize twelve distinct engineering design issues.The algorithms used in the study are listed as:transient search optimization(TSO),equilibrium optimizer(EO),grey wolf optimizer(GWO),moth-flame optimization(MFO),whale optimization algorithm(WOA),slimemould algorithm(SMA),harris hawks optimization(HHO),chimp optimization algorithm(COA),coot optimization algorithm(COOT),multi-verse optimization(MVO),arithmetic optimization algorithm(AOA),aquila optimizer(AO),sine cosine algorithm(SCA),smell agent optimization(SAO),and seagull optimization algorithm(SOA),pelican optimization algorithm(POA),and coati optimization algorithm(CA).As far as we know,there is no comparative analysis of recent and popular methods against the concrete conditions of real-world engineering problems.Hence,a remarkable research guideline is presented in the study for researchersworking in the fields of engineering and artificial intelligence,especiallywhen applying the optimization methods that have emerged recently.Future research can rely on this work for a literature search on comparisons of metaheuristic optimization methods in real-world problems under similar conditions. 展开更多
关键词 Metaheuristic optimization algorithms real-world engineering design problems multidisciplinary design optimization problems
下载PDF
Systematic Review:Load Balancing in Cloud Computing by Using Metaheuristic Based Dynamic Algorithms
11
作者 Darakhshan Syed Ghulam Muhammad Safdar Rizvi 《Intelligent Automation & Soft Computing》 2024年第3期437-476,共40页
Cloud Computing has the ability to provide on-demand access to a shared resource pool.It has completely changed the way businesses are managed,implement applications,and provide services.The rise in popularity has led... Cloud Computing has the ability to provide on-demand access to a shared resource pool.It has completely changed the way businesses are managed,implement applications,and provide services.The rise in popularity has led to a significant increase in the user demand for services.However,in cloud environments efficient load balancing is essential to ensure optimal performance and resource utilization.This systematic review targets a detailed description of load balancing techniques including static and dynamic load balancing algorithms.Specifically,metaheuristic-based dynamic load balancing algorithms are identified as the optimal solution in case of increased traffic.In a cloud-based context,this paper describes load balancing measurements,including the benefits and drawbacks associated with the selected load balancing techniques.It also summarizes the algorithms based on implementation,time complexity,adaptability,associated issue(s),and targeted QoS parameters.Additionally,the analysis evaluates the tools and instruments utilized in each investigated study.Moreover,comparative analysis among static,traditional dynamic and metaheuristic algorithms based on response time by using the CloudSim simulation tool is also performed.Finally,the key open problems and potential directions for the state-of-the-art metaheuristic-based approaches are also addressed. 展开更多
关键词 Cloud computing load balancing metaheuristic algorithm dynamic algorithm load balancer QOS
下载PDF
Decoupling Algorithms for the Gravitational Wave Spacecraft
12
作者 XueWang Weizhou Zhu +4 位作者 Zhao Cui Xingguang Qian Jinke Yang Jianjun Jia Yikun Wang 《Computer Modeling in Engineering & Sciences》 SCIE EI 2024年第7期325-337,共13页
The gravitational wave spacecraft is a complex multi-input multi-output dynamic system.The gravitational wave detection mission requires the spacecraft to achieve single spacecraft with two laser links and high-precis... The gravitational wave spacecraft is a complex multi-input multi-output dynamic system.The gravitational wave detection mission requires the spacecraft to achieve single spacecraft with two laser links and high-precision control.Establishing one spacecraftwith two laser links,compared to one spacecraft with a single laser link,requires an upgraded decoupling algorithmfor the link establishment.The decoupling algorithmwe designed reassigns the degrees of freedomand forces in the control loop to ensure sufficient degrees of freedomfor optical axis control.In addressing the distinct dynamic characteristics of different degrees of freedom,a transfer function compensation method is used in the decoupling process to further minimize motion coupling.The open-loop frequency response of the systemis obtained through simulation.The upgraded decoupling algorithms effectively reduce the open-loop frequency response by 30 dB.The transfer function compensation method efficiently suppresses the coupling of low-frequency noise. 展开更多
关键词 Gravitational waves spacecraft laser acquisition decoupling algorithms dynamical model optical axis control
下载PDF
Quantification of the concrete freeze–thaw environment across the Qinghai–Tibet Plateau based on machine learning algorithms
13
作者 QIN Yanhui MA Haoyuan +3 位作者 ZHANG Lele YIN Jinshuai ZHENG Xionghui LI Shuo 《Journal of Mountain Science》 SCIE CSCD 2024年第1期322-334,共13页
The reasonable quantification of the concrete freezing environment on the Qinghai–Tibet Plateau(QTP) is the primary issue in frost resistant concrete design, which is one of the challenges that the QTP engineering ma... The reasonable quantification of the concrete freezing environment on the Qinghai–Tibet Plateau(QTP) is the primary issue in frost resistant concrete design, which is one of the challenges that the QTP engineering managers should take into account. In this paper, we propose a more realistic method to calculate the number of concrete freeze–thaw cycles(NFTCs) on the QTP. The calculated results show that the NFTCs increase as the altitude of the meteorological station increases with the average NFTCs being 208.7. Four machine learning methods, i.e., the random forest(RF) model, generalized boosting method(GBM), generalized linear model(GLM), and generalized additive model(GAM), are used to fit the NFTCs. The root mean square error(RMSE) values of the RF, GBM, GLM, and GAM are 32.3, 4.3, 247.9, and 161.3, respectively. The R^(2) values of the RF, GBM, GLM, and GAM are 0.93, 0.99, 0.48, and 0.66, respectively. The GBM method performs the best compared to the other three methods, which was shown by the results of RMSE and R^(2) values. The quantitative results from the GBM method indicate that the lowest, medium, and highest NFTC values are distributed in the northern, central, and southern parts of the QTP, respectively. The annual NFTCs in the QTP region are mainly concentrated at 160 and above, and the average NFTCs is 200 across the QTP. Our results can provide scientific guidance and a theoretical basis for the freezing resistance design of concrete in various projects on the QTP. 展开更多
关键词 Freeze–thaw cycles Quantification Machine learning algorithms Qinghai–Tibet Plateau CONCRETE
下载PDF
Accelerating Large-Scale Sorting through Parallel Algorithms
14
作者 Yahya Alhabboub Fares Almutairi +3 位作者 Mohammed Safhi Yazan Alqahtani Adam Almeedani Yasir Alguwaifli 《Journal of Computer and Communications》 2024年第1期131-138,共8页
This study explores the application of parallel algorithms to enhance large-scale sorting, focusing on the QuickSort method. Implemented in both sequential and parallel forms, the paper provides a detailed comparison ... This study explores the application of parallel algorithms to enhance large-scale sorting, focusing on the QuickSort method. Implemented in both sequential and parallel forms, the paper provides a detailed comparison of their performance. This study investigates the efficacy of both techniques through the lens of array generation and pivot selection to manage datasets of varying sizes. This study meticulously documents the performance metrics, recording 16,499.2 milliseconds for the serial implementation and 16,339 milliseconds for the parallel implementation when sorting an array by using C++ chrono library. These results suggest that while the performance gains of the parallel approach over its serial counterpart are not immediately pronounced for smaller datasets, the benefits are expected to be more substantial as the dataset size increases. 展开更多
关键词 Sorting algorithm Quick Sort QuickSort Parallel Parallel algorithms
下载PDF
Enhancing Evolutionary Algorithms With Pattern Mining for Sparse Large-Scale Multi-Objective Optimization Problems
15
作者 Sheng Qi Rui Wang +3 位作者 Tao Zhang Weixiong Huang Fan Yu Ling Wang 《IEEE/CAA Journal of Automatica Sinica》 SCIE EI CSCD 2024年第8期1786-1801,共16页
Sparse large-scale multi-objective optimization problems(SLMOPs)are common in science and engineering.However,the large-scale problem represents the high dimensionality of the decision space,requiring algorithms to tr... Sparse large-scale multi-objective optimization problems(SLMOPs)are common in science and engineering.However,the large-scale problem represents the high dimensionality of the decision space,requiring algorithms to traverse vast expanse with limited computational resources.Furthermore,in the context of sparse,most variables in Pareto optimal solutions are zero,making it difficult for algorithms to identify non-zero variables efficiently.This paper is dedicated to addressing the challenges posed by SLMOPs.To start,we introduce innovative objective functions customized to mine maximum and minimum candidate sets.This substantial enhancement dramatically improves the efficacy of frequent pattern mining.In this way,selecting candidate sets is no longer based on the quantity of nonzero variables they contain but on a higher proportion of nonzero variables within specific dimensions.Additionally,we unveil a novel approach to association rule mining,which delves into the intricate relationships between non-zero variables.This novel methodology aids in identifying sparse distributions that can potentially expedite reductions in the objective function value.We extensively tested our algorithm across eight benchmark problems and four real-world SLMOPs.The results demonstrate that our approach achieves competitive solutions across various challenges. 展开更多
关键词 Evolutionary algorithms pattern mining sparse large-scale multi-objective problems(SLMOPs) sparse large-scale optimization.
下载PDF
Solar Radiation Estimation Based on a New Combined Approach of Artificial Neural Networks (ANN) and Genetic Algorithms (GA) in South Algeria
16
作者 Djeldjli Halima Benatiallah Djelloul +3 位作者 Ghasri Mehdi Tanougast Camel Benatiallah Ali Benabdelkrim Bouchra 《Computers, Materials & Continua》 SCIE EI 2024年第6期4725-4740,共16页
When designing solar systems and assessing the effectiveness of their many uses,estimating sun irradiance is a crucial first step.This study examined three approaches(ANN,GA-ANN,and ANFIS)for estimating daily global s... When designing solar systems and assessing the effectiveness of their many uses,estimating sun irradiance is a crucial first step.This study examined three approaches(ANN,GA-ANN,and ANFIS)for estimating daily global solar radiation(GSR)in the south of Algeria:Adrar,Ouargla,and Bechar.The proposed hybrid GA-ANN model,based on genetic algorithm-based optimization,was developed to improve the ANN model.The GA-ANN and ANFIS models performed better than the standalone ANN-based model,with GA-ANN being better suited for forecasting in all sites,and it performed the best with the best values in the testing phase of Coefficient of Determination(R=0.9005),Mean Absolute Percentage Error(MAPE=8.40%),and Relative Root Mean Square Error(rRMSE=12.56%).Nevertheless,the ANFIS model outperformed the GA-ANN model in forecasting daily GSR,with the best values of indicators when testing the model being R=0.9374,MAPE=7.78%,and rRMSE=10.54%.Generally,we may conclude that the initial ANN stand-alone model performance when forecasting solar radiation has been improved,and the results obtained after injecting the genetic algorithm into the ANN to optimize its weights were satisfactory.The model can be used to forecast daily GSR in dry climates and other climates and may also be helpful in selecting solar energy system installations and sizes. 展开更多
关键词 Solar energy systems genetic algorithm neural networks hybrid adaptive neuro fuzzy inference system solar radiation
下载PDF
Identification of Lubricating Oil Additives Using XGBoost and Ant Colony Optimization Algorithms
17
作者 Xia Yanqiu Cui Jinwei +2 位作者 Xie Peiyuan Zou Shaode Feng Xin 《China Petroleum Processing & Petrochemical Technology》 SCIE CAS CSCD 2024年第2期158-167,共10页
To address the problem of identifying multiple types of additives in lubricating oil,a method based on midinfrared spectral band selection using the eXtreme Gradient Boosting(XGBoost)algorithm combined with the ant co... To address the problem of identifying multiple types of additives in lubricating oil,a method based on midinfrared spectral band selection using the eXtreme Gradient Boosting(XGBoost)algorithm combined with the ant colony optimization(ACO)algorithm is proposed.The XGBoost algorithm was used to train and test three additives,T534(alkyl diphenylamine),T308(isooctyl acid thiophospholipid octadecylamine),and T306(trimethylphenol phosphate),separately,in order to screen for the optimal combination of spectral bands for each additive.The ACO algorithm was used to optimize the parameters of the XGBoost algorithm to improve the identification accuracy.During this process,the support vector machine(SVM)and hybrid bat algorithms(HBA)were included as a comparison,generating four models:ACO-XGBoost,ACO-SVM,HBA-XGboost,and HBA-SVM.The results showed that all four models could identify the three additives efficiently,with the ACO-XGBoost model achieving 100%recognition of all three additives.In addition,the generalizability of the ACO-XGBoost model was further demonstrated by predicting a lubricating oil containing the three additives prepared in our laboratory and a collected sample of commercial oil currently in use。 展开更多
关键词 lubricant oil additives fourier transform infrared spectroscopy type identification ACO-XGBoost combinatorial algorithm
下载PDF
A method for cleaning wind power anomaly data by combining image processing with community detection algorithms
18
作者 Qiaoling Yang Kai Chen +2 位作者 Jianzhang Man Jiaheng Duan Zuoqi Jin 《Global Energy Interconnection》 EI CSCD 2024年第3期293-312,共20页
Current methodologies for cleaning wind power anomaly data exhibit limited capabilities in identifying abnormal data within extensive datasets and struggle to accommodate the considerable variability and intricacy of ... Current methodologies for cleaning wind power anomaly data exhibit limited capabilities in identifying abnormal data within extensive datasets and struggle to accommodate the considerable variability and intricacy of wind farm data.Consequently,a method for cleaning wind power anomaly data by combining image processing with community detection algorithms(CWPAD-IPCDA)is proposed.To precisely identify and initially clean anomalous data,wind power curve(WPC)images are converted into graph structures,which employ the Louvain community recognition algorithm and graph-theoretic methods for community detection and segmentation.Furthermore,the mathematical morphology operation(MMO)determines the main part of the initially cleaned wind power curve images and maps them back to the normal wind power points to complete the final cleaning.The CWPAD-IPCDA method was applied to clean datasets from 25 wind turbines(WTs)in two wind farms in northwest China to validate its feasibility.A comparison was conducted using density-based spatial clustering of applications with noise(DBSCAN)algorithm,an improved isolation forest algorithm,and an image-based(IB)algorithm.The experimental results demonstrate that the CWPAD-IPCDA method surpasses the other three algorithms,achieving an approximately 7.23%higher average data cleaning rate.The mean value of the sum of the squared errors(SSE)of the dataset after cleaning is approximately 6.887 lower than that of the other algorithms.Moreover,the mean of overall accuracy,as measured by the F1-score,exceeds that of the other methods by approximately 10.49%;this indicates that the CWPAD-IPCDA method is more conducive to improving the accuracy and reliability of wind power curve modeling and wind farm power forecasting. 展开更多
关键词 Wind turbine power curve Abnormal data cleaning Community detection Louvain algorithm Mathematical morphology operation
下载PDF
Appropriate Combination of Crossover Operator and Mutation Operator in Genetic Algorithms for the Travelling Salesman Problem
19
作者 Zakir Hussain Ahmed Habibollah Haron Abdullah Al-Tameem 《Computers, Materials & Continua》 SCIE EI 2024年第5期2399-2425,共27页
Genetic algorithms(GAs)are very good metaheuristic algorithms that are suitable for solving NP-hard combinatorial optimization problems.AsimpleGAbeginswith a set of solutions represented by a population of chromosomes... Genetic algorithms(GAs)are very good metaheuristic algorithms that are suitable for solving NP-hard combinatorial optimization problems.AsimpleGAbeginswith a set of solutions represented by a population of chromosomes and then uses the idea of survival of the fittest in the selection process to select some fitter chromosomes.It uses a crossover operator to create better offspring chromosomes and thus,converges the population.Also,it uses a mutation operator to explore the unexplored areas by the crossover operator,and thus,diversifies the GA search space.A combination of crossover and mutation operators makes the GA search strong enough to reach the optimal solution.However,appropriate selection and combination of crossover operator and mutation operator can lead to a very good GA for solving an optimization problem.In this present paper,we aim to study the benchmark traveling salesman problem(TSP).We developed several genetic algorithms using seven crossover operators and six mutation operators for the TSP and then compared them to some benchmark TSPLIB instances.The experimental studies show the effectiveness of the combination of a comprehensive sequential constructive crossover operator and insertion mutation operator for the problem.The GA using the comprehensive sequential constructive crossover with insertion mutation could find average solutions whose average percentage of excesses from the best-known solutions are between 0.22 and 14.94 for our experimented problem instances. 展开更多
关键词 Travelling salesman problem genetic algorithms crossover operator mutation operator comprehensive sequential constructive crossover insertion mutation
下载PDF
Multi-strategy hybrid whale optimization algorithms for complex constrained optimization problems
20
作者 王振宇 WANG Lei 《High Technology Letters》 EI CAS 2024年第1期99-108,共10页
A multi-strategy hybrid whale optimization algorithm(MSHWOA)for complex constrained optimization problems is proposed to overcome the drawbacks of easily trapping into local optimum,slow convergence speed and low opti... A multi-strategy hybrid whale optimization algorithm(MSHWOA)for complex constrained optimization problems is proposed to overcome the drawbacks of easily trapping into local optimum,slow convergence speed and low optimization precision.Firstly,the population is initialized by introducing the theory of good point set,which increases the randomness and diversity of the population and lays the foundation for the global optimization of the algorithm.Then,a novel linearly update equation of convergence factor is designed to coordinate the abilities of exploration and exploitation.At the same time,the global exploration and local exploitation capabilities are improved through the siege mechanism of Harris Hawks optimization algorithm.Finally,the simulation experiments are conducted on the 6 benchmark functions and Wilcoxon rank sum test to evaluate the optimization performance of the improved algorithm.The experimental results show that the proposed algorithm has more significant improvement in optimization accuracy,convergence speed and robustness than the comparison algorithm. 展开更多
关键词 whale optimization algorithm(WOA) good point set nonlinear convergence factor siege mechanism
下载PDF
上一页 1 2 250 下一页 到第
使用帮助 返回顶部