It is one of the topics that have been studied extensively on maximum power point tracking(MPPT)recently.Traditional or soft computing methods are used for MPPT.Since soft computing approaches are more effective than ...It is one of the topics that have been studied extensively on maximum power point tracking(MPPT)recently.Traditional or soft computing methods are used for MPPT.Since soft computing approaches are more effective than traditional approaches,studies on MPPT have shifted in this direction.This study aims comparison of performance of seven meta-heuristic training algorithms in the neuro-fuzzy training for MPPT.The meta-heuristic training algorithms used are particle swarm optimization(PSO),harmony search(HS),cuckoo search(CS),artificial bee colony(ABC)algorithm,bee algorithm(BA),differential evolution(DE)and flower pollination algorithm(FPA).The antecedent and conclusion parameters of neuro-fuzzy are determined by these algorithms.The data of a 250 W photovoltaic(PV)is used in the applications.For effective MPPT,different neuro-fuzzy structures,different membership functions and different control parameter values are evaluated in detail.Related training algorithms are compared in terms of solution quality and convergence speed.The strengths and weaknesses of these algorithms are revealed.It is seen that the type and number of membership function,colony size,number of generations affect the solution quality and convergence speed of the training algorithms.As a result,it has been observed that CS and ABC algorithm are more effective than other algorithms in terms of solution quality and convergence in solving the related problem.展开更多
Damage identification of the offshore floating wind turbine by vibration/dynamic signals is one of the important and new research fields in the Structural Health Monitoring(SHM). In this paper a new damage identific...Damage identification of the offshore floating wind turbine by vibration/dynamic signals is one of the important and new research fields in the Structural Health Monitoring(SHM). In this paper a new damage identification method is proposed based on meta-heuristic algorithms using the dynamic response of the TLP(Tension-Leg Platform) floating wind turbine structure. The Genetic Algorithms(GA), Artificial Immune System(AIS), Particle Swarm Optimization(PSO), and Artificial Bee Colony(ABC) are chosen for minimizing the object function, defined properly for damage identification purpose. In addition to studying the capability of mentioned algorithms in correctly identifying the damage, the effect of the response type on the results of identification is studied. Also, the results of proposed damage identification are investigated with considering possible uncertainties of the structure. Finally, for evaluating the proposed method in real condition, a 1/100 scaled experimental setup of TLP Floating Wind Turbine(TLPFWT) is provided in a laboratory scale and the proposed damage identification method is applied to the scaled turbine.展开更多
Different performance levels may be obtained for sideway collapse evaluation of steel moment frames depending on the evaluation procedure used to handle uncertainties. In this article, the process of representing mode...Different performance levels may be obtained for sideway collapse evaluation of steel moment frames depending on the evaluation procedure used to handle uncertainties. In this article, the process of representing modelling uncertainties, record to record (RTR) variations and cognitive uncertainties for moment resisting steel frames of various heights is discussed in detail. RTR uncertainty is used by incremental dynamic analysis (IDA), modelling uncertainties are considered through backbone curves and hysteresis loops of component, and cognitive uncertainty is presented in three levels of material quality. IDA is used to evaluate RTR uncertainty based on strong ground motion records selected by the k-means algorithm, which is favoured over Monte Carlo selection due to its time saving appeal. Analytical equations of the Response Surface Method are obtained through IDA results by the Cuckoo algorithm, which predicts the mean and standard deviation of the collapse fragility curve. The Takagi-Sugeno-Kang model is used to represent material quality based on the response surface coefficients. Finally, collapse fragility curves with the various sources of uncertainties mentioned are derived through a large number of material quality values and meta variables inferred by the Takagi-Sugeno-Kang fuzzy model based on response surface method coefficients. It is concluded that a better risk management strategy in countries where material quality control is weak, is to account for cognitive uncertainties in fragility curves and the mean annual frequency.展开更多
Small parasitic Hemipteran insects known as bedbugs(Cimicidae)feed on warm-blooded mammal’s blood.The most famous member of this family is the Cimex lectularius or common bedbug.The current paper proposes a novel swa...Small parasitic Hemipteran insects known as bedbugs(Cimicidae)feed on warm-blooded mammal’s blood.The most famous member of this family is the Cimex lectularius or common bedbug.The current paper proposes a novel swarm intelligence optimization algorithm called the Bedbug Meta-Heuristic Algorithm(BMHA).The primary inspiration for the bedbug algorithm comes from the static and dynamic swarming behaviors of bedbugs in nature.The two main stages of optimization algorithms,exploration,and exploitation,are designed by modeling bedbug social interaction to search for food.The proposed algorithm is benchmarked qualitatively and quantitatively using many test functions including CEC2019.The results of evaluating BMHA prove that this algorithm can improve the initial random population for a given optimization problem to converge towards global optimization and provide highly competitive results compared to other well-known optimization algorithms.The results also prove the new algorithm's performance in solving real optimization problems in unknown search spaces.To achieve this,the proposed algorithm has been used to select the features of fake news in a semi-supervised manner,the results of which show the good performance of the proposed algorithm in solving problems.展开更多
Blasting is well-known as an effective method for fragmenting or moving rock in open-pit mines.To evaluate the quality of blasting,the size of rock distribution is used as a critical criterion in blasting operations.A...Blasting is well-known as an effective method for fragmenting or moving rock in open-pit mines.To evaluate the quality of blasting,the size of rock distribution is used as a critical criterion in blasting operations.A high percentage of oversized rocks generated by blasting operations can lead to economic and environmental damage.Therefore,this study proposed four novel intelligent models to predict the size of rock distribution in mine blasting in order to optimize blasting parameters,as well as the efficiency of blasting operation in open mines.Accordingly,a nature-inspired algorithm(i.e.,firefly algorithm-FFA)and different machine learning algorithms(i.e.,gradient boosting machine(GBM),support vector machine(SVM),Gaussian process(GP),and artificial neural network(ANN))were combined for this aim,abbreviated as FFA-GBM,FFA-SVM,FFA-GP,and FFA-ANN,respectively.Subsequently,predicted results from the abovementioned models were compared with each other using three statistical indicators(e.g.,mean absolute error,root-mean-squared error,and correlation coefficient)and color intensity method.For developing and simulating the size of rock in blasting operations,136 blasting events with their images were collected and analyzed by the Split-Desktop software.In which,111 events were randomly selected for the development and optimization of the models.Subsequently,the remaining 25 blasting events were applied to confirm the accuracy of the proposed models.Herein,blast design parameters were regarded as input variables to predict the size of rock in blasting operations.Finally,the obtained results revealed that the FFA is a robust optimization algorithm for estimating rock fragmentation in bench blasting.Among the models developed in this study,FFA-GBM provided the highest accuracy in predicting the size of fragmented rocks.The other techniques(i.e.,FFA-SVM,FFA-GP,and FFA-ANN)yielded lower computational stability and efficiency.Hence,the FFA-GBM model can be used as a powerful and precise soft computing tool that can be applied to practical engineering cases aiming to improve the quality of blasting and rock fragmentation.展开更多
Optimized road maintenance planning seeks for solutions that can minimize the life-cycle cost of a road network and concurrently maximize pavement condition. Aiming at pro- posing an optimal set of road maintenance so...Optimized road maintenance planning seeks for solutions that can minimize the life-cycle cost of a road network and concurrently maximize pavement condition. Aiming at pro- posing an optimal set of road maintenance solutions, robust meta-heuristic algorithms are used in research. Two main optimization techniques are applied including single-objective and multi-objective optimization. Genetic algorithms (GA), particle swarm optimization (PSO), and combination of genetic algorithm and particle swarm optimization (GAPSO) as single-objective techniques are used, while the non-domination sorting genetic algorithm II (NSGAII) and multi-objective particle swarm optimization (MOPSO) which are sufficient for solving computationally complex large-size optimization problems as multi-objective techniques are applied and compared. A real case study from the rural transportation network of Iran is employed to illustrate the sufficiency of the optimum algorithm. The formulation of the optimization model is carried out in such a way that a cost-effective maintenance strategy is reached by preserving the performance level of the road network at a desirable level. So, the objective functions are pavement performance maximization and maintenance cost minimization. It is concluded that multi-objective algorithms including non-domination sorting genetic algorithm II (NSGAII) and multi-objective particle swarm optimization performed better than the single objective algorithms due to the capability to balance between both objectives. And between multi-objective algorithms the NSGAII provides the optimum solution for the road maintenance planning.展开更多
The non-invasive evaluation of the heart through EectroCardioG-raphy(ECG)has played a key role in detecting heart disease.The analysis of ECG signals requires years of learning and experience to interpret and extract ...The non-invasive evaluation of the heart through EectroCardioG-raphy(ECG)has played a key role in detecting heart disease.The analysis of ECG signals requires years of learning and experience to interpret and extract useful information from them.Thus,a computerized system is needed to classify ECG signals with more accurate results effectively.Abnormal heart rhythms are called arrhythmias and cause sudden cardiac deaths.In this work,a Computerized Abnormal Heart Rhythms Detection(CAHRD)system is developed using ECG signals.It consists of four stages;preprocessing,feature extraction,feature optimization and classifier.At first,Pan and Tompkins algorithm is employed to detect the envelope of Q,R and S waves in the preprocessing stage.It uses a recursive filter to eliminate muscle noise,T-wave interference and baseline wander.As the analysis of ECG signal in the spatial domain does not provide a complete description of the signal,the feature extraction involves using frequency contents obtained from multiple wavelet filters;bi-orthogonal,Symlet and Daubechies at different resolution levels in the feature extraction stage.Then,Black Widow Optimization(BWO)is applied to optimize the hybrid wavelet features in the feature optimization stage.Finally,a kernel based Support Vector Machine(SVM)is employed to classify heartbeats into five classes.In SVM,Radial Basis Function(RBF),polynomial and linear kernels are used.A total of∼15000 ECG signals are obtained from the Massachusetts Institute of Technology-Beth Israel Hospital(MIT-BIH)arrhythmia database for performance evaluation of the proposed CAHRD system.Results show that the proposed CAHRD system proved to be a powerful tool for ECG analysis.It correctly classifies five classes of heartbeats with 99.91%accuracy using an RBF kernel with 2nd level wavelet coefficients.The CAHRD system achieves an improvement of∼6%over random projections with the ensemble SVM approach and∼2%over morphological and ECG segment based features with the RBF classifier.展开更多
Compression index Ccis an essential parameter in geotechnical design for which the effectiveness of correlation is still a challenge.This paper suggests a novel modelling approach using machine learning(ML)technique.T...Compression index Ccis an essential parameter in geotechnical design for which the effectiveness of correlation is still a challenge.This paper suggests a novel modelling approach using machine learning(ML)technique.The performance of five commonly used machine learning(ML)algorithms,i.e.back-propagation neural network(BPNN),extreme learning machine(ELM),support vector machine(SVM),random forest(RF)and evolutionary polynomial regression(EPR)in predicting Cc is comprehensively investigated.A database with a total number of 311 datasets including three input variables,i.e.initial void ratio e0,liquid limit water content wL,plasticity index Ip,and one output variable Cc is first established.Genetic algorithm(GA)is used to optimize the hyper-parameters in five ML algorithms,and the average prediction error for the 10-fold cross-validation(CV)sets is set as thefitness function in the GA for enhancing the robustness of ML models.The results indicate that ML models outperform empirical prediction formulations with lower prediction error.RF yields the lowest error followed by BPNN,ELM,EPR and SVM.If the ranges of input variables in the database are large enough,BPNN and RF models are recommended to predict Cc.Furthermore,if the distribution of input variables is continuous,RF model is the best one.Otherwise,EPR model is recommended if the ranges of input variables are small.The predicted correlations between input and output variables using five ML models show great agreement with the physical explanation.展开更多
One of the important research issues in wireless sensor networks(WSNs)is the optimal layout designing for the deployment of sensor nodes.It directly affects the quality of monitoring,cost,and detection capability of W...One of the important research issues in wireless sensor networks(WSNs)is the optimal layout designing for the deployment of sensor nodes.It directly affects the quality of monitoring,cost,and detection capability of WSNs.Layout optimization is an NP-hard combinatorial problem,which requires optimization of multiple competing objectives like cost,coverage,connectivity,lifetime,load balancing,and energy consumption of sensor nodes.In the last decade,several meta-heuristic optimization techniques have been proposed to solve this problem,such as genetic algorithms(GA)and particle swarm optimization(PSO).However,these approaches either provided computationally expensive solutions or covered a limited number of objectives,which are combinations of area coverage,the number of sensor nodes,energy consumption,and lifetime.In this study,a meta-heuristic multi-objective firefly algorithm(MOFA)is presented to solve the layout optimization problem.Here,the main goal is to cover a number of objectives related to optimal layouts of homogeneous WSNs,which includes coverage,connectivity,lifetime,energy consumption and the number of sensor nodes.Simulation results showed that MOFA created optimal Pareto front of non-dominated solutions with better hyper-volumes and spread of solutions,in comparison to multi-objective genetic algorithms(IBEA,NSGA-II)and particle swarm optimizers(OMOPSO,SMOPSO).Therefore,MOFA can be used in real-time deployment applications of large-scale WSNs to enhance their detection capability and quality of monitoring.展开更多
Soil swelling-related disaster is considered as one of the most devastating geo-hazards in modern history.Hence,proper determination of a soil’s ability to expand is very vital for achieving a secure and safe ground ...Soil swelling-related disaster is considered as one of the most devastating geo-hazards in modern history.Hence,proper determination of a soil’s ability to expand is very vital for achieving a secure and safe ground for infrastructures.Accordingly,this study has provided a novel and intelligent approach that enables an improved estimation of swelling by using kernelised machines(Bayesian linear regression(BLR)&bayes point machine(BPM)support vector machine(SVM)and deep-support vector machine(D-SVM));(multiple linear regressor(REG),logistic regressor(LR)and artificial neural network(ANN)),tree-based algorithms such as decision forest(RDF)&boosted trees(BDT).Also,and for the first time,meta-heuristic classifiers incorporating the techniques of voting(VE)and stacking(SE)were utilised.Different independent scenarios of explanatory features’combination that influence soil behaviour in swelling were investigated.Preliminary results indicated BLR as possessing the highest amount of deviation from the predictor variable(the actual swell-strain).REG and BLR performed slightly better than ANN while the meta-heuristic learners(VE and SE)produced the best overall performance(greatest R2 value of 0.94 and RMSE of 0.06%exhibited by VE).CEC,plasticity index and moisture content were the features considered to have the highest level of importance.Kernelized binary classifiers(SVM,D-SVM and BPM)gave better accuracy(average accuracy and recall rate of 0.93 and 0.60)compared to ANN,LR and RDF.Sensitivity-driven diagnostic test indicated that the meta-heuristic models’best performance occurred when ML training was conducted using k-fold validation technique.Finally,it is recommended that the concepts developed herein be deployed during the preliminary phases of a geotechnical or geological site characterisation by using the best performing meta-heuristic models via their background coding resource.展开更多
In this study, we propose an algorithm selection method based on coupling strength for the partitioned analysis ofstructure-piezoelectric-circuit coupling, which includes two types of coupling or inverse and direct pi...In this study, we propose an algorithm selection method based on coupling strength for the partitioned analysis ofstructure-piezoelectric-circuit coupling, which includes two types of coupling or inverse and direct piezoelectriccoupling and direct piezoelectric and circuit coupling. In the proposed method, implicit and explicit formulationsare used for strong and weak coupling, respectively. Three feasible partitioned algorithms are generated, namely(1) a strongly coupled algorithm that uses a fully implicit formulation for both types of coupling, (2) a weaklycoupled algorithm that uses a fully explicit formulation for both types of coupling, and (3) a partially stronglycoupled and partially weakly coupled algorithm that uses an implicit formulation and an explicit formulation forthe two types of coupling, respectively.Numerical examples using a piezoelectric energy harvester,which is a typicalstructure-piezoelectric-circuit coupling problem, demonstrate that the proposed method selects the most costeffectivealgorithm.展开更多
The Gannet Optimization Algorithm (GOA) and the Whale Optimization Algorithm (WOA) demonstrate strong performance;however, there remains room for improvement in convergence and practical applications. This study intro...The Gannet Optimization Algorithm (GOA) and the Whale Optimization Algorithm (WOA) demonstrate strong performance;however, there remains room for improvement in convergence and practical applications. This study introduces a hybrid optimization algorithm, named the adaptive inertia weight whale optimization algorithm and gannet optimization algorithm (AIWGOA), which addresses challenges in enhancing handwritten documents. The hybrid strategy integrates the strengths of both algorithms, significantly enhancing their capabilities, whereas the adaptive parameter strategy mitigates the need for manual parameter setting. By amalgamating the hybrid strategy and parameter-adaptive approach, the Gannet Optimization Algorithm was refined to yield the AIWGOA. Through a performance analysis of the CEC2013 benchmark, the AIWGOA demonstrates notable advantages across various metrics. Subsequently, an evaluation index was employed to assess the enhanced handwritten documents and images, affirming the superior practical application of the AIWGOA compared with other algorithms.展开更多
A Rapid-exploration Random Tree(RRT)autonomous detection algorithm based on the multi-guide-node deflection strategy and Karto Simultaneous Localization and Mapping(SLAM)algorithm was proposed to solve the problems of...A Rapid-exploration Random Tree(RRT)autonomous detection algorithm based on the multi-guide-node deflection strategy and Karto Simultaneous Localization and Mapping(SLAM)algorithm was proposed to solve the problems of low efficiency of detecting frontier boundary points and drift distortion in the process of map building in the traditional RRT algorithm in the autonomous detection strategy of mobile robot.Firstly,an RRT global frontier boundary point detection algorithm based on the multi-guide-node deflection strategy was put forward,which introduces the reference value of guide nodes’deflection probability into the random sampling function so that the global search tree can detect frontier boundary points towards the guide nodes according to random probability.After that,a new autonomous detection algorithm for mobile robots was proposed by combining the graph optimization-based Karto SLAM algorithm with the previously improved RRT algorithm.The algorithm simulation platform based on the Gazebo platform was built.The simulation results show that compared with the traditional RRT algorithm,the proposed RRT autonomous detection algorithm can effectively reduce the time of autonomous detection,plan the length of detection trajectory under the condition of high average detection coverage,and complete the task of autonomous detection mapping more efficiently.Finally,with the help of the ROS-based mobile robot experimental platform,the performance of the proposed algorithm was verified in the real environment of different obstacles.The experimental results show that in the actual environment of simple and complex obstacles,the proposed RRT autonomous detection algorithm was superior to the traditional RRT autonomous detection algorithm in the time of detection,length of detection trajectory,and average coverage,thus improving the efficiency and accuracy of autonomous detection.展开更多
Blank holder force(BHF)is a crucial parameter in deep drawing,having close relation with the forming quality of sheet metal.However,there are different BHFs maintaining the best forming effect in different stages of d...Blank holder force(BHF)is a crucial parameter in deep drawing,having close relation with the forming quality of sheet metal.However,there are different BHFs maintaining the best forming effect in different stages of deep drawing.The variable blank holder force(VBHF)varying with the drawing stage can overcome this problem at an extent.The optimization of VBHF is to determine the optimal BHF in every deep drawing stage.In this paper,a new heuristic optimization algorithm named Jaya is introduced to solve the optimization efficiently.An improved“Quasi-oppositional”strategy is added to Jaya algorithm for improving population diversity.Meanwhile,an innovated stop criterion is added for better convergence.Firstly,the quality evaluation criteria for wrinkling and tearing are built.Secondly,the Kriging models are developed to approximate and quantify the relation between VBHF and forming defects under random sampling.Finally,the optimization models are established and solved by the improved QO-Jaya algorithm.A VBHF optimization example of component with complicated shape and thin wall is studied to prove the effectiveness of the improved Jaya algorithm.The optimization results are compared with that obtained by other algorithms based on the TOPSIS method.展开更多
This research paper presents a novel optimization method called the Synergistic Swarm Optimization Algorithm(SSOA).The SSOA combines the principles of swarmintelligence and synergistic cooperation to search for optima...This research paper presents a novel optimization method called the Synergistic Swarm Optimization Algorithm(SSOA).The SSOA combines the principles of swarmintelligence and synergistic cooperation to search for optimal solutions efficiently.A synergistic cooperation mechanism is employed,where particles exchange information and learn from each other to improve their search behaviors.This cooperation enhances the exploitation of promising regions in the search space while maintaining exploration capabilities.Furthermore,adaptive mechanisms,such as dynamic parameter adjustment and diversification strategies,are incorporated to balance exploration and exploitation.By leveraging the collaborative nature of swarm intelligence and integrating synergistic cooperation,the SSOAmethod aims to achieve superior convergence speed and solution quality performance compared to other optimization algorithms.The effectiveness of the proposed SSOA is investigated in solving the 23 benchmark functions and various engineering design problems.The experimental results highlight the effectiveness and potential of the SSOA method in addressing challenging optimization problems,making it a promising tool for a wide range of applications in engineering and beyond.Matlab codes of SSOA are available at:https://www.mathworks.com/matlabcentral/fileexchange/153466-synergistic-swarm-optimization-algorithm.展开更多
Cloud computing provides a diverse and adaptable resource pool over the internet,allowing users to tap into various resources as needed.It has been seen as a robust solution to relevant challenges.A significant delay ...Cloud computing provides a diverse and adaptable resource pool over the internet,allowing users to tap into various resources as needed.It has been seen as a robust solution to relevant challenges.A significant delay can hamper the performance of IoT-enabled cloud platforms.However,efficient task scheduling can lower the cloud infrastructure’s energy consumption,thus maximizing the service provider’s revenue by decreasing user job processing times.The proposed Modified Chimp-Whale Optimization Algorithm called Modified Chimp-Whale Optimization Algorithm(MCWOA),combines elements of the Chimp Optimization Algorithm(COA)and the Whale Optimization Algorithm(WOA).To enhance MCWOA’s identification precision,the Sobol sequence is used in the population initialization phase,ensuring an even distribution of the population across the solution space.Moreover,the traditional MCWOA’s local search capabilities are augmented by incorporating the whale optimization algorithm’s bubble-net hunting and random search mechanisms into MCWOA’s position-updating process.This study demonstrates the effectiveness of the proposed approach using a two-story rigid frame and a simply supported beam model.Simulated outcomes reveal that the new method outperforms the original MCWOA,especially in multi-damage detection scenarios.MCWOA excels in avoiding false positives and enhancing computational speed,making it an optimal choice for structural damage detection.The efficiency of the proposed MCWOA is assessed against metrics such as energy usage,computational expense,task duration,and delay.The simulated data indicates that the new MCWOA outpaces other methods across all metrics.The study also references the Whale Optimization Algorithm(WOA),Chimp Algorithm(CA),Ant Lion Optimizer(ALO),Genetic Algorithm(GA)and Grey Wolf Optimizer(GWO).展开更多
Localization or positioning scheme in Wireless sensor networks (WSNs) is one of the most challenging andfundamental operations in various monitoring or tracking applications because the network deploys a large areaand...Localization or positioning scheme in Wireless sensor networks (WSNs) is one of the most challenging andfundamental operations in various monitoring or tracking applications because the network deploys a large areaand allocates the acquired location information to unknown devices. The metaheuristic approach is one of themost advantageous ways to deal with this challenging issue and overcome the disadvantages of the traditionalmethods that often suffer from computational time problems and small network deployment scale. This studyproposes an enhanced whale optimization algorithm that is an advanced metaheuristic algorithm based on thesiege mechanism (SWOA) for node localization inWSN. The objective function is modeled while communicatingon localized nodes, considering variables like delay, path loss, energy, and received signal strength. The localizationapproach also assigns the discovered location data to unidentified devices with the modeled objective functionby applying the SWOA algorithm. The experimental analysis is carried out to demonstrate the efficiency of thedesigned localization scheme in terms of various metrics, e.g., localization errors rate, converges rate, and executedtime. Compared experimental-result shows that theSWOA offers the applicability of the developed model forWSNto perform the localization scheme with excellent quality. Significantly, the error and convergence values achievedby the SWOA are less location error, faster in convergence and executed time than the others compared to at least areduced 1.5% to 4.7% error rate, and quicker by at least 4%and 2% in convergence and executed time, respectivelyfor the experimental scenarios.展开更多
In view of the complex marine environment of navigation,especially in the case of multiple static and dynamic obstacles,the traditional obstacle avoidance algorithms applied to unmanned surface vehicles(USV)are prone ...In view of the complex marine environment of navigation,especially in the case of multiple static and dynamic obstacles,the traditional obstacle avoidance algorithms applied to unmanned surface vehicles(USV)are prone to fall into the trap of local optimization.Therefore,this paper proposes an improved artificial potential field(APF)algorithm,which uses 5G communication technology to communicate between the USV and the control center.The algorithm introduces the USV discrimination mechanism to avoid the USV falling into local optimization when the USV encounter different obstacles in different scenarios.Considering the various scenarios between the USV and other dynamic obstacles such as vessels in the process of performing tasks,the algorithm introduces the concept of dynamic artificial potential field.For the multiple obstacles encountered in the process of USV sailing,based on the International Regulations for Preventing Collisions at Sea(COLREGS),the USV determines whether the next step will fall into local optimization through the discriminationmechanism.The local potential field of the USV will dynamically adjust,and the reverse virtual gravitational potential field will be added to prevent it from falling into the local optimization and avoid collisions.The objective function and cost function are designed at the same time,so that the USV can smoothly switch between the global path and the local obstacle avoidance.The simulation results show that the improved APF algorithm proposed in this paper can successfully avoid various obstacles in the complex marine environment,and take navigation time and economic cost into account.展开更多
In the cloud environment,ensuring a high level of data security is in high demand.Data planning storage optimization is part of the whole security process in the cloud environment.It enables data security by avoiding ...In the cloud environment,ensuring a high level of data security is in high demand.Data planning storage optimization is part of the whole security process in the cloud environment.It enables data security by avoiding the risk of data loss and data overlapping.The development of data flow scheduling approaches in the cloud environment taking security parameters into account is insufficient.In our work,we propose a data scheduling model for the cloud environment.Themodel is made up of three parts that together help dispatch user data flow to the appropriate cloudVMs.The first component is the Collector Agent whichmust periodically collect information on the state of the network links.The second one is the monitoring agent which must then analyze,classify,and make a decision on the state of the link and finally transmit this information to the scheduler.The third one is the scheduler who must consider previous information to transfer user data,including fair distribution and reliable paths.It should be noted that each part of the proposedmodel requires the development of its algorithms.In this article,we are interested in the development of data transfer algorithms,including fairness distribution with the consideration of a stable link state.These algorithms are based on the grouping of transmitted files and the iterative method.The proposed algorithms showthe performances to obtain an approximate solution to the studied problem which is an NP-hard(Non-Polynomial solution)problem.The experimental results show that the best algorithm is the half-grouped minimum excluding(HME),with a percentage of 91.3%,an average deviation of 0.042,and an execution time of 0.001 s.展开更多
Blockchain technology,with its attributes of decentralization,immutability,and traceability,has emerged as a powerful catalyst for enhancing traditional industries in terms of optimizing business processes.However,tra...Blockchain technology,with its attributes of decentralization,immutability,and traceability,has emerged as a powerful catalyst for enhancing traditional industries in terms of optimizing business processes.However,transaction performance and scalability has become the main challenges hindering the widespread adoption of blockchain.Due to its inability to meet the demands of high-frequency trading,blockchain cannot be adopted in many scenarios.To improve the transaction capacity,researchers have proposed some on-chain scaling technologies,including lightning networks,directed acyclic graph technology,state channels,and shardingmechanisms,inwhich sharding emerges as a potential scaling technology.Nevertheless,excessive cross-shard transactions and uneven shard workloads prevent the sharding mechanism from achieving the expected aim.This paper proposes a graphbased sharding scheme for public blockchain to efficiently balance the transaction distribution.Bymitigating crossshard transactions and evening-out workloads among shards,the scheme reduces transaction confirmation latency and enhances the transaction capacity of the blockchain.Therefore,the scheme can achieve a high-frequency transaction as well as a better blockchain scalability.Experiments results show that the scheme effectively reduces the cross-shard transaction ratio to a range of 35%-56%and significantly decreases the transaction confirmation latency to 6 s in a blockchain with no more than 25 shards.展开更多
文摘It is one of the topics that have been studied extensively on maximum power point tracking(MPPT)recently.Traditional or soft computing methods are used for MPPT.Since soft computing approaches are more effective than traditional approaches,studies on MPPT have shifted in this direction.This study aims comparison of performance of seven meta-heuristic training algorithms in the neuro-fuzzy training for MPPT.The meta-heuristic training algorithms used are particle swarm optimization(PSO),harmony search(HS),cuckoo search(CS),artificial bee colony(ABC)algorithm,bee algorithm(BA),differential evolution(DE)and flower pollination algorithm(FPA).The antecedent and conclusion parameters of neuro-fuzzy are determined by these algorithms.The data of a 250 W photovoltaic(PV)is used in the applications.For effective MPPT,different neuro-fuzzy structures,different membership functions and different control parameter values are evaluated in detail.Related training algorithms are compared in terms of solution quality and convergence speed.The strengths and weaknesses of these algorithms are revealed.It is seen that the type and number of membership function,colony size,number of generations affect the solution quality and convergence speed of the training algorithms.As a result,it has been observed that CS and ABC algorithm are more effective than other algorithms in terms of solution quality and convergence in solving the related problem.
文摘Damage identification of the offshore floating wind turbine by vibration/dynamic signals is one of the important and new research fields in the Structural Health Monitoring(SHM). In this paper a new damage identification method is proposed based on meta-heuristic algorithms using the dynamic response of the TLP(Tension-Leg Platform) floating wind turbine structure. The Genetic Algorithms(GA), Artificial Immune System(AIS), Particle Swarm Optimization(PSO), and Artificial Bee Colony(ABC) are chosen for minimizing the object function, defined properly for damage identification purpose. In addition to studying the capability of mentioned algorithms in correctly identifying the damage, the effect of the response type on the results of identification is studied. Also, the results of proposed damage identification are investigated with considering possible uncertainties of the structure. Finally, for evaluating the proposed method in real condition, a 1/100 scaled experimental setup of TLP Floating Wind Turbine(TLPFWT) is provided in a laboratory scale and the proposed damage identification method is applied to the scaled turbine.
文摘Different performance levels may be obtained for sideway collapse evaluation of steel moment frames depending on the evaluation procedure used to handle uncertainties. In this article, the process of representing modelling uncertainties, record to record (RTR) variations and cognitive uncertainties for moment resisting steel frames of various heights is discussed in detail. RTR uncertainty is used by incremental dynamic analysis (IDA), modelling uncertainties are considered through backbone curves and hysteresis loops of component, and cognitive uncertainty is presented in three levels of material quality. IDA is used to evaluate RTR uncertainty based on strong ground motion records selected by the k-means algorithm, which is favoured over Monte Carlo selection due to its time saving appeal. Analytical equations of the Response Surface Method are obtained through IDA results by the Cuckoo algorithm, which predicts the mean and standard deviation of the collapse fragility curve. The Takagi-Sugeno-Kang model is used to represent material quality based on the response surface coefficients. Finally, collapse fragility curves with the various sources of uncertainties mentioned are derived through a large number of material quality values and meta variables inferred by the Takagi-Sugeno-Kang fuzzy model based on response surface method coefficients. It is concluded that a better risk management strategy in countries where material quality control is weak, is to account for cognitive uncertainties in fragility curves and the mean annual frequency.
文摘Small parasitic Hemipteran insects known as bedbugs(Cimicidae)feed on warm-blooded mammal’s blood.The most famous member of this family is the Cimex lectularius or common bedbug.The current paper proposes a novel swarm intelligence optimization algorithm called the Bedbug Meta-Heuristic Algorithm(BMHA).The primary inspiration for the bedbug algorithm comes from the static and dynamic swarming behaviors of bedbugs in nature.The two main stages of optimization algorithms,exploration,and exploitation,are designed by modeling bedbug social interaction to search for food.The proposed algorithm is benchmarked qualitatively and quantitatively using many test functions including CEC2019.The results of evaluating BMHA prove that this algorithm can improve the initial random population for a given optimization problem to converge towards global optimization and provide highly competitive results compared to other well-known optimization algorithms.The results also prove the new algorithm's performance in solving real optimization problems in unknown search spaces.To achieve this,the proposed algorithm has been used to select the features of fake news in a semi-supervised manner,the results of which show the good performance of the proposed algorithm in solving problems.
基金supported by the Center for Mining,Electro-Mechanical research of Hanoi University of Mining and Geology(HUMG),Hanoi,Vietnamfinancially supported by the Hunan Provincial Department of Education General Project(19C1744)+1 种基金Hunan Province Science Foundation for Youth Scholars of China fund(2018JJ3510)the Innovation-Driven Project of Central South University(2020CX040)。
文摘Blasting is well-known as an effective method for fragmenting or moving rock in open-pit mines.To evaluate the quality of blasting,the size of rock distribution is used as a critical criterion in blasting operations.A high percentage of oversized rocks generated by blasting operations can lead to economic and environmental damage.Therefore,this study proposed four novel intelligent models to predict the size of rock distribution in mine blasting in order to optimize blasting parameters,as well as the efficiency of blasting operation in open mines.Accordingly,a nature-inspired algorithm(i.e.,firefly algorithm-FFA)and different machine learning algorithms(i.e.,gradient boosting machine(GBM),support vector machine(SVM),Gaussian process(GP),and artificial neural network(ANN))were combined for this aim,abbreviated as FFA-GBM,FFA-SVM,FFA-GP,and FFA-ANN,respectively.Subsequently,predicted results from the abovementioned models were compared with each other using three statistical indicators(e.g.,mean absolute error,root-mean-squared error,and correlation coefficient)and color intensity method.For developing and simulating the size of rock in blasting operations,136 blasting events with their images were collected and analyzed by the Split-Desktop software.In which,111 events were randomly selected for the development and optimization of the models.Subsequently,the remaining 25 blasting events were applied to confirm the accuracy of the proposed models.Herein,blast design parameters were regarded as input variables to predict the size of rock in blasting operations.Finally,the obtained results revealed that the FFA is a robust optimization algorithm for estimating rock fragmentation in bench blasting.Among the models developed in this study,FFA-GBM provided the highest accuracy in predicting the size of fragmented rocks.The other techniques(i.e.,FFA-SVM,FFA-GP,and FFA-ANN)yielded lower computational stability and efficiency.Hence,the FFA-GBM model can be used as a powerful and precise soft computing tool that can be applied to practical engineering cases aiming to improve the quality of blasting and rock fragmentation.
文摘Optimized road maintenance planning seeks for solutions that can minimize the life-cycle cost of a road network and concurrently maximize pavement condition. Aiming at pro- posing an optimal set of road maintenance solutions, robust meta-heuristic algorithms are used in research. Two main optimization techniques are applied including single-objective and multi-objective optimization. Genetic algorithms (GA), particle swarm optimization (PSO), and combination of genetic algorithm and particle swarm optimization (GAPSO) as single-objective techniques are used, while the non-domination sorting genetic algorithm II (NSGAII) and multi-objective particle swarm optimization (MOPSO) which are sufficient for solving computationally complex large-size optimization problems as multi-objective techniques are applied and compared. A real case study from the rural transportation network of Iran is employed to illustrate the sufficiency of the optimum algorithm. The formulation of the optimization model is carried out in such a way that a cost-effective maintenance strategy is reached by preserving the performance level of the road network at a desirable level. So, the objective functions are pavement performance maximization and maintenance cost minimization. It is concluded that multi-objective algorithms including non-domination sorting genetic algorithm II (NSGAII) and multi-objective particle swarm optimization performed better than the single objective algorithms due to the capability to balance between both objectives. And between multi-objective algorithms the NSGAII provides the optimum solution for the road maintenance planning.
文摘The non-invasive evaluation of the heart through EectroCardioG-raphy(ECG)has played a key role in detecting heart disease.The analysis of ECG signals requires years of learning and experience to interpret and extract useful information from them.Thus,a computerized system is needed to classify ECG signals with more accurate results effectively.Abnormal heart rhythms are called arrhythmias and cause sudden cardiac deaths.In this work,a Computerized Abnormal Heart Rhythms Detection(CAHRD)system is developed using ECG signals.It consists of four stages;preprocessing,feature extraction,feature optimization and classifier.At first,Pan and Tompkins algorithm is employed to detect the envelope of Q,R and S waves in the preprocessing stage.It uses a recursive filter to eliminate muscle noise,T-wave interference and baseline wander.As the analysis of ECG signal in the spatial domain does not provide a complete description of the signal,the feature extraction involves using frequency contents obtained from multiple wavelet filters;bi-orthogonal,Symlet and Daubechies at different resolution levels in the feature extraction stage.Then,Black Widow Optimization(BWO)is applied to optimize the hybrid wavelet features in the feature optimization stage.Finally,a kernel based Support Vector Machine(SVM)is employed to classify heartbeats into five classes.In SVM,Radial Basis Function(RBF),polynomial and linear kernels are used.A total of∼15000 ECG signals are obtained from the Massachusetts Institute of Technology-Beth Israel Hospital(MIT-BIH)arrhythmia database for performance evaluation of the proposed CAHRD system.Results show that the proposed CAHRD system proved to be a powerful tool for ECG analysis.It correctly classifies five classes of heartbeats with 99.91%accuracy using an RBF kernel with 2nd level wavelet coefficients.The CAHRD system achieves an improvement of∼6%over random projections with the ensemble SVM approach and∼2%over morphological and ECG segment based features with the RBF classifier.
基金financial support provided by the RIF project(Grant No.PolyU R5037-18F)from the Research Grants Council(RGC)of Hong Kong is gratefully acknowledged。
文摘Compression index Ccis an essential parameter in geotechnical design for which the effectiveness of correlation is still a challenge.This paper suggests a novel modelling approach using machine learning(ML)technique.The performance of five commonly used machine learning(ML)algorithms,i.e.back-propagation neural network(BPNN),extreme learning machine(ELM),support vector machine(SVM),random forest(RF)and evolutionary polynomial regression(EPR)in predicting Cc is comprehensively investigated.A database with a total number of 311 datasets including three input variables,i.e.initial void ratio e0,liquid limit water content wL,plasticity index Ip,and one output variable Cc is first established.Genetic algorithm(GA)is used to optimize the hyper-parameters in five ML algorithms,and the average prediction error for the 10-fold cross-validation(CV)sets is set as thefitness function in the GA for enhancing the robustness of ML models.The results indicate that ML models outperform empirical prediction formulations with lower prediction error.RF yields the lowest error followed by BPNN,ELM,EPR and SVM.If the ranges of input variables in the database are large enough,BPNN and RF models are recommended to predict Cc.Furthermore,if the distribution of input variables is continuous,RF model is the best one.Otherwise,EPR model is recommended if the ranges of input variables are small.The predicted correlations between input and output variables using five ML models show great agreement with the physical explanation.
基金This research has been funded by the Deanship of Scientific Research at Imam Mohammad Ibn Saud Islamic University through Research Group No.RG-21-07-09.
文摘One of the important research issues in wireless sensor networks(WSNs)is the optimal layout designing for the deployment of sensor nodes.It directly affects the quality of monitoring,cost,and detection capability of WSNs.Layout optimization is an NP-hard combinatorial problem,which requires optimization of multiple competing objectives like cost,coverage,connectivity,lifetime,load balancing,and energy consumption of sensor nodes.In the last decade,several meta-heuristic optimization techniques have been proposed to solve this problem,such as genetic algorithms(GA)and particle swarm optimization(PSO).However,these approaches either provided computationally expensive solutions or covered a limited number of objectives,which are combinations of area coverage,the number of sensor nodes,energy consumption,and lifetime.In this study,a meta-heuristic multi-objective firefly algorithm(MOFA)is presented to solve the layout optimization problem.Here,the main goal is to cover a number of objectives related to optimal layouts of homogeneous WSNs,which includes coverage,connectivity,lifetime,energy consumption and the number of sensor nodes.Simulation results showed that MOFA created optimal Pareto front of non-dominated solutions with better hyper-volumes and spread of solutions,in comparison to multi-objective genetic algorithms(IBEA,NSGA-II)and particle swarm optimizers(OMOPSO,SMOPSO).Therefore,MOFA can be used in real-time deployment applications of large-scale WSNs to enhance their detection capability and quality of monitoring.
文摘Soil swelling-related disaster is considered as one of the most devastating geo-hazards in modern history.Hence,proper determination of a soil’s ability to expand is very vital for achieving a secure and safe ground for infrastructures.Accordingly,this study has provided a novel and intelligent approach that enables an improved estimation of swelling by using kernelised machines(Bayesian linear regression(BLR)&bayes point machine(BPM)support vector machine(SVM)and deep-support vector machine(D-SVM));(multiple linear regressor(REG),logistic regressor(LR)and artificial neural network(ANN)),tree-based algorithms such as decision forest(RDF)&boosted trees(BDT).Also,and for the first time,meta-heuristic classifiers incorporating the techniques of voting(VE)and stacking(SE)were utilised.Different independent scenarios of explanatory features’combination that influence soil behaviour in swelling were investigated.Preliminary results indicated BLR as possessing the highest amount of deviation from the predictor variable(the actual swell-strain).REG and BLR performed slightly better than ANN while the meta-heuristic learners(VE and SE)produced the best overall performance(greatest R2 value of 0.94 and RMSE of 0.06%exhibited by VE).CEC,plasticity index and moisture content were the features considered to have the highest level of importance.Kernelized binary classifiers(SVM,D-SVM and BPM)gave better accuracy(average accuracy and recall rate of 0.93 and 0.60)compared to ANN,LR and RDF.Sensitivity-driven diagnostic test indicated that the meta-heuristic models’best performance occurred when ML training was conducted using k-fold validation technique.Finally,it is recommended that the concepts developed herein be deployed during the preliminary phases of a geotechnical or geological site characterisation by using the best performing meta-heuristic models via their background coding resource.
基金the Japan Society for the Promotion of Science,KAKENHI Grant Nos.20H04199 and 23H00475.
文摘In this study, we propose an algorithm selection method based on coupling strength for the partitioned analysis ofstructure-piezoelectric-circuit coupling, which includes two types of coupling or inverse and direct piezoelectriccoupling and direct piezoelectric and circuit coupling. In the proposed method, implicit and explicit formulationsare used for strong and weak coupling, respectively. Three feasible partitioned algorithms are generated, namely(1) a strongly coupled algorithm that uses a fully implicit formulation for both types of coupling, (2) a weaklycoupled algorithm that uses a fully explicit formulation for both types of coupling, and (3) a partially stronglycoupled and partially weakly coupled algorithm that uses an implicit formulation and an explicit formulation forthe two types of coupling, respectively.Numerical examples using a piezoelectric energy harvester,which is a typicalstructure-piezoelectric-circuit coupling problem, demonstrate that the proposed method selects the most costeffectivealgorithm.
文摘The Gannet Optimization Algorithm (GOA) and the Whale Optimization Algorithm (WOA) demonstrate strong performance;however, there remains room for improvement in convergence and practical applications. This study introduces a hybrid optimization algorithm, named the adaptive inertia weight whale optimization algorithm and gannet optimization algorithm (AIWGOA), which addresses challenges in enhancing handwritten documents. The hybrid strategy integrates the strengths of both algorithms, significantly enhancing their capabilities, whereas the adaptive parameter strategy mitigates the need for manual parameter setting. By amalgamating the hybrid strategy and parameter-adaptive approach, the Gannet Optimization Algorithm was refined to yield the AIWGOA. Through a performance analysis of the CEC2013 benchmark, the AIWGOA demonstrates notable advantages across various metrics. Subsequently, an evaluation index was employed to assess the enhanced handwritten documents and images, affirming the superior practical application of the AIWGOA compared with other algorithms.
基金This research was funded by National Natural Science Foundation of China(No.62063006)Guangxi Science and Technology Major Program(No.2022AA05002)+2 种基金Key Laboratory of AI and Information Processing(Hechi University),Education Department of Guangxi Zhuang Autonomous Region(No.2022GXZDSY003)Guangxi Key Laboratory of Spatial Information and Geomatics(Guilin University of Technology)(No.21-238-21-16)Innovation Project of Guangxi Graduate Education(No.YCSW2023352).
文摘A Rapid-exploration Random Tree(RRT)autonomous detection algorithm based on the multi-guide-node deflection strategy and Karto Simultaneous Localization and Mapping(SLAM)algorithm was proposed to solve the problems of low efficiency of detecting frontier boundary points and drift distortion in the process of map building in the traditional RRT algorithm in the autonomous detection strategy of mobile robot.Firstly,an RRT global frontier boundary point detection algorithm based on the multi-guide-node deflection strategy was put forward,which introduces the reference value of guide nodes’deflection probability into the random sampling function so that the global search tree can detect frontier boundary points towards the guide nodes according to random probability.After that,a new autonomous detection algorithm for mobile robots was proposed by combining the graph optimization-based Karto SLAM algorithm with the previously improved RRT algorithm.The algorithm simulation platform based on the Gazebo platform was built.The simulation results show that compared with the traditional RRT algorithm,the proposed RRT autonomous detection algorithm can effectively reduce the time of autonomous detection,plan the length of detection trajectory under the condition of high average detection coverage,and complete the task of autonomous detection mapping more efficiently.Finally,with the help of the ROS-based mobile robot experimental platform,the performance of the proposed algorithm was verified in the real environment of different obstacles.The experimental results show that in the actual environment of simple and complex obstacles,the proposed RRT autonomous detection algorithm was superior to the traditional RRT autonomous detection algorithm in the time of detection,length of detection trajectory,and average coverage,thus improving the efficiency and accuracy of autonomous detection.
基金Supported by National Key Research and Development Program of China(Grant No.2022YFB3304200)National Natural Science Foundation of China(Grant No.52075479)Taizhou Municipal Science and Technology Project of China(Grant No.1801gy23).
文摘Blank holder force(BHF)is a crucial parameter in deep drawing,having close relation with the forming quality of sheet metal.However,there are different BHFs maintaining the best forming effect in different stages of deep drawing.The variable blank holder force(VBHF)varying with the drawing stage can overcome this problem at an extent.The optimization of VBHF is to determine the optimal BHF in every deep drawing stage.In this paper,a new heuristic optimization algorithm named Jaya is introduced to solve the optimization efficiently.An improved“Quasi-oppositional”strategy is added to Jaya algorithm for improving population diversity.Meanwhile,an innovated stop criterion is added for better convergence.Firstly,the quality evaluation criteria for wrinkling and tearing are built.Secondly,the Kriging models are developed to approximate and quantify the relation between VBHF and forming defects under random sampling.Finally,the optimization models are established and solved by the improved QO-Jaya algorithm.A VBHF optimization example of component with complicated shape and thin wall is studied to prove the effectiveness of the improved Jaya algorithm.The optimization results are compared with that obtained by other algorithms based on the TOPSIS method.
基金King Saud University for funding this research through Researchers Supporting Program Number(RSPD2023R704),King Saud University,Riyadh,Saudi Arabia.
文摘This research paper presents a novel optimization method called the Synergistic Swarm Optimization Algorithm(SSOA).The SSOA combines the principles of swarmintelligence and synergistic cooperation to search for optimal solutions efficiently.A synergistic cooperation mechanism is employed,where particles exchange information and learn from each other to improve their search behaviors.This cooperation enhances the exploitation of promising regions in the search space while maintaining exploration capabilities.Furthermore,adaptive mechanisms,such as dynamic parameter adjustment and diversification strategies,are incorporated to balance exploration and exploitation.By leveraging the collaborative nature of swarm intelligence and integrating synergistic cooperation,the SSOAmethod aims to achieve superior convergence speed and solution quality performance compared to other optimization algorithms.The effectiveness of the proposed SSOA is investigated in solving the 23 benchmark functions and various engineering design problems.The experimental results highlight the effectiveness and potential of the SSOA method in addressing challenging optimization problems,making it a promising tool for a wide range of applications in engineering and beyond.Matlab codes of SSOA are available at:https://www.mathworks.com/matlabcentral/fileexchange/153466-synergistic-swarm-optimization-algorithm.
文摘Cloud computing provides a diverse and adaptable resource pool over the internet,allowing users to tap into various resources as needed.It has been seen as a robust solution to relevant challenges.A significant delay can hamper the performance of IoT-enabled cloud platforms.However,efficient task scheduling can lower the cloud infrastructure’s energy consumption,thus maximizing the service provider’s revenue by decreasing user job processing times.The proposed Modified Chimp-Whale Optimization Algorithm called Modified Chimp-Whale Optimization Algorithm(MCWOA),combines elements of the Chimp Optimization Algorithm(COA)and the Whale Optimization Algorithm(WOA).To enhance MCWOA’s identification precision,the Sobol sequence is used in the population initialization phase,ensuring an even distribution of the population across the solution space.Moreover,the traditional MCWOA’s local search capabilities are augmented by incorporating the whale optimization algorithm’s bubble-net hunting and random search mechanisms into MCWOA’s position-updating process.This study demonstrates the effectiveness of the proposed approach using a two-story rigid frame and a simply supported beam model.Simulated outcomes reveal that the new method outperforms the original MCWOA,especially in multi-damage detection scenarios.MCWOA excels in avoiding false positives and enhancing computational speed,making it an optimal choice for structural damage detection.The efficiency of the proposed MCWOA is assessed against metrics such as energy usage,computational expense,task duration,and delay.The simulated data indicates that the new MCWOA outpaces other methods across all metrics.The study also references the Whale Optimization Algorithm(WOA),Chimp Algorithm(CA),Ant Lion Optimizer(ALO),Genetic Algorithm(GA)and Grey Wolf Optimizer(GWO).
基金the VNUHCM-University of Information Technology’s Scientific Research Support Fund.
文摘Localization or positioning scheme in Wireless sensor networks (WSNs) is one of the most challenging andfundamental operations in various monitoring or tracking applications because the network deploys a large areaand allocates the acquired location information to unknown devices. The metaheuristic approach is one of themost advantageous ways to deal with this challenging issue and overcome the disadvantages of the traditionalmethods that often suffer from computational time problems and small network deployment scale. This studyproposes an enhanced whale optimization algorithm that is an advanced metaheuristic algorithm based on thesiege mechanism (SWOA) for node localization inWSN. The objective function is modeled while communicatingon localized nodes, considering variables like delay, path loss, energy, and received signal strength. The localizationapproach also assigns the discovered location data to unidentified devices with the modeled objective functionby applying the SWOA algorithm. The experimental analysis is carried out to demonstrate the efficiency of thedesigned localization scheme in terms of various metrics, e.g., localization errors rate, converges rate, and executedtime. Compared experimental-result shows that theSWOA offers the applicability of the developed model forWSNto perform the localization scheme with excellent quality. Significantly, the error and convergence values achievedby the SWOA are less location error, faster in convergence and executed time than the others compared to at least areduced 1.5% to 4.7% error rate, and quicker by at least 4%and 2% in convergence and executed time, respectivelyfor the experimental scenarios.
基金This work was supported by the Postdoctoral Fund of FDCT,Macao(Grant No.0003/2021/APD).Any opinions,findings and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect those of the sponsor.
文摘In view of the complex marine environment of navigation,especially in the case of multiple static and dynamic obstacles,the traditional obstacle avoidance algorithms applied to unmanned surface vehicles(USV)are prone to fall into the trap of local optimization.Therefore,this paper proposes an improved artificial potential field(APF)algorithm,which uses 5G communication technology to communicate between the USV and the control center.The algorithm introduces the USV discrimination mechanism to avoid the USV falling into local optimization when the USV encounter different obstacles in different scenarios.Considering the various scenarios between the USV and other dynamic obstacles such as vessels in the process of performing tasks,the algorithm introduces the concept of dynamic artificial potential field.For the multiple obstacles encountered in the process of USV sailing,based on the International Regulations for Preventing Collisions at Sea(COLREGS),the USV determines whether the next step will fall into local optimization through the discriminationmechanism.The local potential field of the USV will dynamically adjust,and the reverse virtual gravitational potential field will be added to prevent it from falling into the local optimization and avoid collisions.The objective function and cost function are designed at the same time,so that the USV can smoothly switch between the global path and the local obstacle avoidance.The simulation results show that the improved APF algorithm proposed in this paper can successfully avoid various obstacles in the complex marine environment,and take navigation time and economic cost into account.
基金the deputyship for Research&Innovation,Ministry of Education in Saudi Arabia for funding this research work through the Project Number(IFP-2022-34).
文摘In the cloud environment,ensuring a high level of data security is in high demand.Data planning storage optimization is part of the whole security process in the cloud environment.It enables data security by avoiding the risk of data loss and data overlapping.The development of data flow scheduling approaches in the cloud environment taking security parameters into account is insufficient.In our work,we propose a data scheduling model for the cloud environment.Themodel is made up of three parts that together help dispatch user data flow to the appropriate cloudVMs.The first component is the Collector Agent whichmust periodically collect information on the state of the network links.The second one is the monitoring agent which must then analyze,classify,and make a decision on the state of the link and finally transmit this information to the scheduler.The third one is the scheduler who must consider previous information to transfer user data,including fair distribution and reliable paths.It should be noted that each part of the proposedmodel requires the development of its algorithms.In this article,we are interested in the development of data transfer algorithms,including fairness distribution with the consideration of a stable link state.These algorithms are based on the grouping of transmitted files and the iterative method.The proposed algorithms showthe performances to obtain an approximate solution to the studied problem which is an NP-hard(Non-Polynomial solution)problem.The experimental results show that the best algorithm is the half-grouped minimum excluding(HME),with a percentage of 91.3%,an average deviation of 0.042,and an execution time of 0.001 s.
基金supported by Shandong Provincial Key Research and Development Program of China(2021CXGC010107,2020CXGC010107)the Shandong Provincial Natural Science Foundation of China(ZR2020KF035)the New 20 Project of Higher Education of Jinan,China(202228017).
文摘Blockchain technology,with its attributes of decentralization,immutability,and traceability,has emerged as a powerful catalyst for enhancing traditional industries in terms of optimizing business processes.However,transaction performance and scalability has become the main challenges hindering the widespread adoption of blockchain.Due to its inability to meet the demands of high-frequency trading,blockchain cannot be adopted in many scenarios.To improve the transaction capacity,researchers have proposed some on-chain scaling technologies,including lightning networks,directed acyclic graph technology,state channels,and shardingmechanisms,inwhich sharding emerges as a potential scaling technology.Nevertheless,excessive cross-shard transactions and uneven shard workloads prevent the sharding mechanism from achieving the expected aim.This paper proposes a graphbased sharding scheme for public blockchain to efficiently balance the transaction distribution.Bymitigating crossshard transactions and evening-out workloads among shards,the scheme reduces transaction confirmation latency and enhances the transaction capacity of the blockchain.Therefore,the scheme can achieve a high-frequency transaction as well as a better blockchain scalability.Experiments results show that the scheme effectively reduces the cross-shard transaction ratio to a range of 35%-56%and significantly decreases the transaction confirmation latency to 6 s in a blockchain with no more than 25 shards.