This paper proposes an autonomous maneuver decision method using transfer learning pigeon-inspired optimization(TLPIO)for unmanned combat aerial vehicles(UCAVs)in dogfight engagements.Firstly,a nonlinear F-16 aircraft...This paper proposes an autonomous maneuver decision method using transfer learning pigeon-inspired optimization(TLPIO)for unmanned combat aerial vehicles(UCAVs)in dogfight engagements.Firstly,a nonlinear F-16 aircraft model and automatic control system are constructed by a MATLAB/Simulink platform.Secondly,a 3-degrees-of-freedom(3-DOF)aircraft model is used as a maneuvering command generator,and the expanded elemental maneuver library is designed,so that the aircraft state reachable set can be obtained.Then,the game matrix is composed with the air combat situation evaluation function calculated according to the angle and range threats.Finally,a key point is that the objective function to be optimized is designed using the game mixed strategy,and the optimal mixed strategy is obtained by TLPIO.Significantly,the proposed TLPIO does not initialize the population randomly,but adopts the transfer learning method based on Kullback-Leibler(KL)divergence to initialize the population,which improves the search accuracy of the optimization algorithm.Besides,the convergence and time complexity of TLPIO are discussed.Comparison analysis with other classical optimization algorithms highlights the advantage of TLPIO.In the simulation of air combat,three initial scenarios are set,namely,opposite,offensive and defensive conditions.The effectiveness performance of the proposed autonomous maneuver decision method is verified by simulation results.展开更多
Currently,the improvement in AI is mainly related to deep learning techniques that are employed for the classification,identification,and quantification of patterns in clinical images.The deep learning models show mor...Currently,the improvement in AI is mainly related to deep learning techniques that are employed for the classification,identification,and quantification of patterns in clinical images.The deep learning models show more remarkable performance than the traditional methods for medical image processing tasks,such as skin cancer,colorectal cancer,brain tumour,cardiac disease,Breast cancer(BrC),and a few more.The manual diagnosis of medical issues always requires an expert and is also expensive.Therefore,developing some computer diagnosis techniques based on deep learning is essential.Breast cancer is the most frequently diagnosed cancer in females with a rapidly growing percentage.It is estimated that patients with BrC will rise to 70%in the next 20 years.If diagnosed at a later stage,the survival rate of patients with BrC is shallow.Hence,early detection is essential,increasing the survival rate to 50%.A new framework for BrC classification is presented that utilises deep learning and feature optimization.The significant steps of the presented framework include(i)hybrid contrast enhancement of acquired images,(ii)data augmentation to facilitate better learning of the Convolutional Neural Network(CNN)model,(iii)a pre‐trained ResNet‐101 model is utilised and modified according to selected dataset classes,(iv)deep transfer learning based model training for feature extraction,(v)the fusion of features using the proposed highly corrected function‐controlled canonical correlation analysis approach,and(vi)optimal feature selection using the modified Satin Bowerbird Optimization controlled Newton Raphson algorithm that finally classified using 10 machine learning classifiers.The experiments of the proposed framework have been carried out using the most critical and publicly available dataset,such as CBISDDSM,and obtained the best accuracy of 94.5%along with improved computation time.The comparison depicts that the presented method surpasses the current state‐ofthe‐art approaches.展开更多
Lung cancer is among the most frequent cancers in the world,with over one million deaths per year.Classification is required for lung cancer diagnosis and therapy to be effective,accurate,and reliable.Gene expression ...Lung cancer is among the most frequent cancers in the world,with over one million deaths per year.Classification is required for lung cancer diagnosis and therapy to be effective,accurate,and reliable.Gene expression microarrays have made it possible to find genetic biomarkers for cancer diagnosis and prediction in a high-throughput manner.Machine Learning(ML)has been widely used to diagnose and classify lung cancer where the performance of ML methods is evaluated to identify the appropriate technique.Identifying and selecting the gene expression patterns can help in lung cancer diagnoses and classification.Normally,microarrays include several genes and may cause confusion or false prediction.Therefore,the Arithmetic Optimization Algorithm(AOA)is used to identify the optimal gene subset to reduce the number of selected genes.Which can allow the classifiers to yield the best performance for lung cancer classification.In addition,we proposed a modified version of AOA which can work effectively on the high dimensional dataset.In the modified AOA,the features are ranked by their weights and are used to initialize the AOA population.The exploitation process of AOA is then enhanced by developing a local search algorithm based on two neighborhood strategies.Finally,the efficiency of the proposed methods was evaluated on gene expression datasets related to Lung cancer using stratified 4-fold cross-validation.The method’s efficacy in selecting the optimal gene subset is underscored by its ability to maintain feature proportions between 10%to 25%.Moreover,the approach significantly enhances lung cancer prediction accuracy.For instance,Lung_Harvard1 achieved an accuracy of 97.5%,Lung_Harvard2 and Lung_Michigan datasets both achieved 100%,Lung_Adenocarcinoma obtained an accuracy of 88.2%,and Lung_Ontario achieved an accuracy of 87.5%.In conclusion,the results indicate the potential promise of the proposed modified AOA approach in classifying microarray cancer data.展开更多
For training the present Neural Network(NN)models,the standard technique is to utilize decaying Learning Rates(LR).While the majority of these techniques commence with a large LR,they will decay multiple times over ti...For training the present Neural Network(NN)models,the standard technique is to utilize decaying Learning Rates(LR).While the majority of these techniques commence with a large LR,they will decay multiple times over time.Decaying has been proved to enhance generalization as well as optimization.Other parameters,such as the network’s size,the number of hidden layers,drop-outs to avoid overfitting,batch size,and so on,are solely based on heuristics.This work has proposed Adaptive Teaching Learning Based(ATLB)Heuristic to identify the optimal hyperparameters for diverse networks.Here we consider three architec-tures Recurrent Neural Networks(RNN),Long Short Term Memory(LSTM),Bidirectional Long Short Term Memory(BiLSTM)of Deep Neural Networks for classification.The evaluation of the proposed ATLB is done through the various learning rate schedulers Cyclical Learning Rate(CLR),Hyperbolic Tangent Decay(HTD),and Toggle between Hyperbolic Tangent Decay and Triangular mode with Restarts(T-HTR)techniques.Experimental results have shown the performance improvement on the 20Newsgroup,Reuters Newswire and IMDB dataset.展开更多
Low-rank matrix recovery is an important problem extensively studied in machine learning, data mining and computer vision communities. A novel method is proposed for low-rank matrix recovery, targeting at higher recov...Low-rank matrix recovery is an important problem extensively studied in machine learning, data mining and computer vision communities. A novel method is proposed for low-rank matrix recovery, targeting at higher recovery accuracy and stronger theoretical guarantee. Specifically, the proposed method is based on a nonconvex optimization model, by solving the low-rank matrix which can be recovered from the noisy observation. To solve the model, an effective algorithm is derived by minimizing over the variables alternately. It is proved theoretically that this algorithm has stronger theoretical guarantee than the existing work. In natural image denoising experiments, the proposed method achieves lower recovery error than the two compared methods. The proposed low-rank matrix recovery method is also applied to solve two real-world problems, i.e., removing noise from verification code and removing watermark from images, in which the images recovered by the proposed method are less noisy than those of the two compared methods.展开更多
In Internet of Things (IoT), large amount of data are processed andcommunicated through different network technologies. Wireless Body Area Networks (WBAN) plays pivotal role in the health care domain with an integrat...In Internet of Things (IoT), large amount of data are processed andcommunicated through different network technologies. Wireless Body Area Networks (WBAN) plays pivotal role in the health care domain with an integration ofIoT and Artificial Intelligence (AI). The amalgamation of above mentioned toolshas taken the new peak in terms of diagnosis and treatment process especially inthe pandemic period. But the real challenges such as low latency, energy consumption high throughput still remains in the dark side of the research. This paperproposes a novel optimized cognitive learning based BAN model based on FogIoT technology as a real-time health monitoring systems with the increased network-life time. Energy and latency aware features of BAN have been extractedand used to train the proposed fog based learning algorithm to achieve low energyconsumption and low-latency scheduling algorithm. To test the proposed network,Fog-IoT-BAN test bed has been developed with the battery driven MICOTTboards interfaced with the health care sensors using Micro Python programming.The extensive experimentation is carried out using the above test beds and variousparameters such as accuracy, precision, recall, F1score and specificity has beencalculated along with QoS (quality of service) parameters such as latency, energyand throughput. To prove the superiority of the proposed framework, the performance of the proposed learning based framework has been compared with theother state-of-art classical learning frameworks and other existing Fog-BAN networks such as WORN, DARE, L-No-DEAF networks. Results proves the proposed framework has outperformed the other classical learning models in termsof accuracy and high False Alarm Rate (FAR), energy efficiency and latency.展开更多
Millimeter wave communication works in the 30–300 GHz frequency range,and can obtain a very high bandwidth,which greatly improves the transmission rate of the communication system and becomes one of the key technolog...Millimeter wave communication works in the 30–300 GHz frequency range,and can obtain a very high bandwidth,which greatly improves the transmission rate of the communication system and becomes one of the key technologies of fifth-generation(5G).The smaller wavelength of the millimeter wave makes it possible to assemble a large number of antennas in a small aperture.The resulting array gain can compensate for the path loss of the millimeter wave.Utilizing this feature,the millimeter wave massive multiple-input multiple-output(MIMO)system uses a large antenna array at the base station.It enables the transmission of multiple data streams,making the system have a higher data transmission rate.In the millimeter wave massive MIMO system,the precoding technology uses the state information of the channel to adjust the transmission strategy at the transmitting end,and the receiving end performs equalization,so that users can better obtain the antenna multiplexing gain and improve the system capacity.This paper proposes an efficient algorithm based on machine learning(ML)for effective system performance in mmwave massive MIMO systems.The main idea is to optimize the adaptive connection structure to maximize the received signal power of each user and correlate the RF chain and base station antenna.Simulation results show that,the proposed algorithm effectively improved the system performance in terms of spectral efficiency and complexity as compared with existing algorithms.展开更多
A chest radiology scan can significantly aid the early diagnosis and management of COVID-19 since the virus attacks the lungs.Chest X-ray(CXR)gained much interest after the COVID-19 outbreak thanks to its rapid imagin...A chest radiology scan can significantly aid the early diagnosis and management of COVID-19 since the virus attacks the lungs.Chest X-ray(CXR)gained much interest after the COVID-19 outbreak thanks to its rapid imaging time,widespread availability,low cost,and portability.In radiological investigations,computer-aided diagnostic tools are implemented to reduce intra-and inter-observer variability.Using lately industrialized Artificial Intelligence(AI)algorithms and radiological techniques to diagnose and classify disease is advantageous.The current study develops an automatic identification and classification model for CXR pictures using Gaussian Fil-tering based Optimized Synergic Deep Learning using Remora Optimization Algorithm(GF-OSDL-ROA).This method is inclusive of preprocessing and classification based on optimization.The data is preprocessed using Gaussian filtering(GF)to remove any extraneous noise from the image’s edges.Then,the OSDL model is applied to classify the CXRs under different severity levels based on CXR data.The learning rate of OSDL is optimized with the help of ROA for COVID-19 diagnosis showing the novelty of the work.OSDL model,applied in this study,was validated using the COVID-19 dataset.The experiments were conducted upon the proposed OSDL model,which achieved a classification accuracy of 99.83%,while the current Convolutional Neural Network achieved less classification accuracy,i.e.,98.14%.展开更多
The exponential growth of Internet of Things(IoT)and 5G networks has resulted in maximum users,and the role of cognitive radio has become pivotal in handling the crowded users.In this scenario,cognitive radio techniqu...The exponential growth of Internet of Things(IoT)and 5G networks has resulted in maximum users,and the role of cognitive radio has become pivotal in handling the crowded users.In this scenario,cognitive radio techniques such as spectrum sensing,spectrum sharing and dynamic spectrum access will become essential components in Wireless IoT communication.IoT devices must learn adaptively to the environment and extract the spectrum knowledge and inferred spectrum knowledge by appropriately changing communication parameters such as modulation index,frequency bands,coding rate etc.,to accommodate the above characteristics.Implementing the above learning methods on the embedded chip leads to high latency,high power consumption and more chip area utilisation.To overcome the problems mentioned above,we present DEEP HOLE Radio sys-tems,the intelligent system enabling the spectrum knowledge extraction from the unprocessed samples by the optimized deep learning models directly from the Radio Frequency(RF)environment.DEEP HOLE Radio provides(i)an opti-mized deep learning framework with a good trade-off between latency,power and utilization.(ii)Complete Hardware-Software architecture where the SoC’s coupled with radio transceivers for maximum performance.The experimentation has been carried out using GNURADIO software interfaced with Zynq-7000 devices mounting on ESP8266 radio transceivers with inbuilt Omni direc-tional antennas.The whole spectrum of knowledge has been extracted using GNU radio.These extracted features are used to train the proposed optimized deep learning models,which run parallel on Zynq-SoC 7000,consuming less area,power,latency and less utilization area.The proposed framework has been evaluated and compared with the existing frameworks such as RFLearn,Long Term Short Memory(LSTM),Convolutional Neural Networks(CNN)and Deep Neural Networks(DNN).The outcome shows that the proposed framework has outperformed the existing framework regarding the area,power and time.More-over,the experimental results show that the proposed framework decreases the delay,power and area by 15%,20%25%concerning the existing RFlearn and other hardware constraint frameworks.展开更多
Surface coating is a critical procedure in the case of maintenance engineering. Ceramic coating of the wear areas is of the best practice which substantially enhances the Mean Time between Failure (MTBF). EN24 is a co...Surface coating is a critical procedure in the case of maintenance engineering. Ceramic coating of the wear areas is of the best practice which substantially enhances the Mean Time between Failure (MTBF). EN24 is a commercial grade alloy which is used for various industrial applications like sleeves, nuts, bolts, shafts, etc. EN24 is having comparatively low corrosion resistance, and ceramic coating of the wear and corroding areas of such parts is a best followed practice which highly improves the frequent failures. The coating quality mainly depends on the coating thickness, surface roughness and coating hardness which finally decides the operability. This paper describes an experimental investigation to effectively optimize the Atmospheric Plasma Spray process input parameters of Al<sub>2</sub>O<sub>3</sub>-40% TiO<sub>2</sub> coatings to get the best quality of coating on EN24 alloy steel substrate. The experiments are conducted with an Orthogonal Array (OA) design of experiments (DoE). In the current experiment, critical input parameters are considered and some of the vital output parameters are monitored accordingly and separate mathematical models are generated using regression analysis. The Analytic Hierarchy Process (AHP) method is used to generate weights for the individual objective functions and based on that, a combined objective function is made. An advanced optimization method, Teaching-Learning-Based Optimization algorithm (TLBO), is practically utilized to the combined objective function to optimize the values of input parameters to get the best output parameters. Confirmation tests are also conducted and their output results are compared with predicted values obtained through mathematical models. The dominating effects of Al<sub>2</sub>O<sub>3</sub>-40% TiO<sub>2</sub> spray parameters on output parameters: surface roughness, coating thickness and coating hardness are discussed in detail. It is concluded that the input parameters variation directly affects the characteristics of output parameters and any number of input as well as output parameters can be easily optimized using the current approach.展开更多
SS304 is a commercial grade stainless steel which is used for various engineering applications like shafts, guides, jigs, fixtures, etc. Ceramic coating of the wear areas of such parts is a regular practice which sign...SS304 is a commercial grade stainless steel which is used for various engineering applications like shafts, guides, jigs, fixtures, etc. Ceramic coating of the wear areas of such parts is a regular practice which significantly enhances the Mean Time Between Failure (MTBF). The final coating quality depends mainly on the coating thickness, surface roughness and hardness which ultimately decides the life. This paper presents an experimental study to effectively optimize the Atmospheric Plasma Spray (APS) process input parameters of Al<sub>2</sub>O<sub>3</sub>-40% TiO2 ceramic coatings to get the best quality of coating on commercial SS304 substrate. The experiments are conducted with a three-level L<sub>18</sub> Orthogonal Array (OA) Design of Experiments (DoE). Critical input parameters considered are: spray nozzle distance, substrate rotating speed, current of the arc, carrier gas flow and coating powder flow rate. The surface roughness, coating thickness and hardness are considered as the output parameters. Mathematical models are generated using regression analysis for individual output parameters. The Analytic Hierarchy Process (AHP) method is applied to generate weights for the individual objective functions and a combined objective function is generated. An advanced optimization method, Teaching-Learning-Based Optimization algorithm (TLBO), is applied to the combined objective function to optimize the values of input parameters to get the best output parameters and confirmation tests are conducted based on that. The significant effects of spray parameters on surface roughness, coating thickness and coating hardness are studied in detail.展开更多
Learning to optimize(L2O)stands at the intersection of traditional optimization and machine learning,utilizing the capabilities of machine learning to enhance conventional optimization techniques.As real-world optimiz...Learning to optimize(L2O)stands at the intersection of traditional optimization and machine learning,utilizing the capabilities of machine learning to enhance conventional optimization techniques.As real-world optimization problems frequently share common structures,L2O provides a tool to exploit these structures for better or faster solutions.This tutorial dives deep into L2O techniques,introducing how to accelerate optimization algorithms,promptly estimate the solutions,or even reshape the optimization problem itself,making it more adaptive to real-world applications.By considering the prerequisites for successful applications of L2O and the structure of the optimization problems at hand,this tutorial provides a comprehensive guide for practitioners and researchers alike.展开更多
Phishing attacks present a persistent and evolving threat in the cybersecurity land-scape,necessitating the development of more sophisticated detection methods.Traditional machine learning approaches to phishing detec...Phishing attacks present a persistent and evolving threat in the cybersecurity land-scape,necessitating the development of more sophisticated detection methods.Traditional machine learning approaches to phishing detection have relied heavily on feature engineering and have often fallen short in adapting to the dynamically changing patterns of phishingUniformResource Locator(URLs).Addressing these challenge,we introduce a framework that integrates the sequential data processing strengths of a Recurrent Neural Network(RNN)with the hyperparameter optimization prowess of theWhale Optimization Algorithm(WOA).Ourmodel capitalizes on an extensive Kaggle dataset,featuring over 11,000 URLs,each delineated by 30 attributes.The WOA’s hyperparameter optimization enhances the RNN’s performance,evidenced by a meticulous validation process.The results,encapsulated in precision,recall,and F1-score metrics,surpass baseline models,achieving an overall accuracy of 92%.This study not only demonstrates the RNN’s proficiency in learning complex patterns but also underscores the WOA’s effectiveness in refining machine learning models for the critical task of phishing detection.展开更多
The query processing in distributed database management systems(DBMS)faces more challenges,such as more operators,and more factors in cost models and meta-data,than that in a single-node DMBS,in which query optimizati...The query processing in distributed database management systems(DBMS)faces more challenges,such as more operators,and more factors in cost models and meta-data,than that in a single-node DMBS,in which query optimization is already an NP-hard problem.Learned query optimizers(mainly in the single-node DBMS)receive attention due to its capability to capture data distributions and flexible ways to avoid hard-craft rules in refinement and adaptation to new hardware.In this paper,we focus on extensions of learned query optimizers to distributed DBMSs.Specifically,we propose one possible but general architecture of the learned query optimizer in the distributed context and highlight differences from the learned optimizer in the single-node ones.In addition,we discuss the challenges and possible solutions.展开更多
Purpose: This study aimed to enhance the prediction of container dwell time, a crucial factor for optimizing port operations, resource allocation, and supply chain efficiency. Determining an optimal learning rate for ...Purpose: This study aimed to enhance the prediction of container dwell time, a crucial factor for optimizing port operations, resource allocation, and supply chain efficiency. Determining an optimal learning rate for training Artificial Neural Networks (ANNs) has remained a challenging task due to the diverse sizes, complexity, and types of data involved. Design/Method/Approach: This research used a RandomizedSearchCV algorithm, a random search approach, to bridge this knowledge gap. The algorithm was applied to container dwell time data from the TOS system of the Port of Tema, which included 307,594 container records from 2014 to 2022. Findings: The RandomizedSearchCV method outperformed standard training methods both in terms of reducing training time and improving prediction accuracy, highlighting the significant role of the constant learning rate as a hyperparameter. Research Limitations and Implications: Although the study provides promising outcomes, the results are limited to the data extracted from the Port of Tema and may differ in other contexts. Further research is needed to generalize these findings across various port systems. Originality/Value: This research underscores the potential of RandomizedSearchCV as a valuable tool for optimizing ANN training in container dwell time prediction. It also accentuates the significance of automated learning rate selection, offering novel insights into the optimization of container dwell time prediction, with implications for improving port efficiency and supply chain operations.展开更多
For unmanned aerial vehicle(UAV)swarm dynamic combat,swarm antagonistic motion control and attack target allocation are extremely challenging sub-tasks.In this paper,the competitive learning pigeon-inspired optimizati...For unmanned aerial vehicle(UAV)swarm dynamic combat,swarm antagonistic motion control and attack target allocation are extremely challenging sub-tasks.In this paper,the competitive learning pigeon-inspired optimization(CLPIO)algorithm is proposed to handle the cooperative dynamic combat problem,which integrates the distributed swarm antagonistic motion and centralized attack target allocation.Moreover,the threshold trigger strategy is presented to switch two sub-tasks.To seek a feasible and optimal combat scheme,a dynamic game approach combined with hawk grouping mechanism and situation assessment between sub-groups is designed to guide the solution of the optimal attack scheme,and the model of swarm antagonistic motion imitating pigeon’s intelligence is proposed to form a confrontation situation.The analysis of the CLPIO algorithm shows its convergence in theory and the comparison with the other four metaheuristic algorithms shows its superiority in solving the mixed Nash equilibrium problem.Finally,numerical simulation verifis that the proposed methods can provide an effective combat scheme in the set scenario.展开更多
We propose multi-objective social learning pigeon-inspired optimization(MSLPIO)and apply it to obstacle avoidance for unmanned aerial vehicle(UAV)formation.In the algorithm,each pigeon learns from the better pigeon bu...We propose multi-objective social learning pigeon-inspired optimization(MSLPIO)and apply it to obstacle avoidance for unmanned aerial vehicle(UAV)formation.In the algorithm,each pigeon learns from the better pigeon but not necessarily the global best one in the update process.A social learning factor is added to the map and compass operator and the landmark operator.In addition,a dimension-dependent parameter setting method is adopted to improve the blindness of parameter setting.We simulate the flight process of five UAVs in a complex obstacle environment.Results verify the effectiveness of the proposed method.MSLPIO has better convergence performance compared with the improved multi-objective pigeon-inspired optimization and the improved non-dominated sorting genetic algorithm.展开更多
This paper employs an efficacious analytical tool,adaptive simplified human learning optimization(ASHLO)algorithm,to solve optimal power flow(OPF)problem in AC/DC hybrid power system,considering valve-point loading ef...This paper employs an efficacious analytical tool,adaptive simplified human learning optimization(ASHLO)algorithm,to solve optimal power flow(OPF)problem in AC/DC hybrid power system,considering valve-point loading effects of generators,carbon tax,and prohibited operating zones of generators,respectively.ASHLO algorithm,involves random learning operator,individual learning operator,social learning operator and adaptive strategies.To compare and analyze the computation performance of the ASHLO method,the proposed ASHLO method and other heuristic intelligent optimization methods are employed to solve OPF problem on the modified IEEE 30-bus and 118-bus AC/DC hybrid test system.Numerical results indicate that the ASHLO method has good convergent property and robustness.Meanwhile,the impacts of wind speeds and locations of HVDC transmission line integrated into the AC network on the OPF results are systematically analyzed.展开更多
The hybrid flow shop scheduling problem with unrelated parallel machine is a typical NP-hard combinatorial optimization problem, and it exists widely in chemical, manufacturing and pharmaceutical industry. In this wor...The hybrid flow shop scheduling problem with unrelated parallel machine is a typical NP-hard combinatorial optimization problem, and it exists widely in chemical, manufacturing and pharmaceutical industry. In this work, a novel mathematic model for the hybrid flow shop scheduling problem with unrelated parallel machine(HFSPUPM) was proposed. Additionally, an effective hybrid estimation of distribution algorithm was proposed to solve the HFSPUPM, taking advantage of the features in the mathematic model. In the optimization algorithm, a new individual representation method was adopted. The(EDA) structure was used for global search while the teaching learning based optimization(TLBO) strategy was used for local search. Based on the structure of the HFSPUPM, this work presents a series of discrete operations. Simulation results show the effectiveness of the proposed hybrid algorithm compared with other algorithms.展开更多
To maximize the power density of the electric propulsion motor in aerospace application,this paper proposes a novel Dynamic Neighborhood Genetic Learning Particle Swarm Optimization(DNGL-PSO)for the motor design,which...To maximize the power density of the electric propulsion motor in aerospace application,this paper proposes a novel Dynamic Neighborhood Genetic Learning Particle Swarm Optimization(DNGL-PSO)for the motor design,which can deal with the insufficient population diversity and non-global optimal solution issues.The DNGL-PSO framework is composed of the dynamic neighborhood module and the particle update module.To improve the population diversity,the dynamic neighborhood strategy is first proposed,which combines the local neighborhood exemplar generation mechanism and the shuffling mechanism.The local neighborhood exemplar generation mechanism enlarges the search range of the algorithm in the solution space,thus obtaining highquality exemplars.Meanwhile,when the global optimal solution cannot update its fitness value,the shuffling mechanism module is triggered to dynamically change the local neighborhood members.The roulette wheel selection operator is introduced into the shuffling mechanism to ensure that particles with larger fitness value are selected with a higher probability and remain in the local neighborhood.Then,the global learning based particle update approach is proposed,which can achieve a good balance between the expansion of the search range in the early stage and the acceleration of local convergence in the later stage.Finally,the optimization design of the electric propulsion motor is conducted to verify the effectiveness of the proposed DNGL-PSO.The simulation results show that the proposed DNGL-PSO has excellent adaptability,optimization efficiency and global optimization capability,while the optimized electric propulsion motor has a high power density of 5.207 kW/kg with the efficiency of 96.12%.展开更多
基金the Science and Technology Innovation 2030-Key Project of“New Generation Artificial Intelligence”(2018AAA0100803)the National Natural Science Foundation of China(U20B2071,91948204,T2121003,U1913602)。
文摘This paper proposes an autonomous maneuver decision method using transfer learning pigeon-inspired optimization(TLPIO)for unmanned combat aerial vehicles(UCAVs)in dogfight engagements.Firstly,a nonlinear F-16 aircraft model and automatic control system are constructed by a MATLAB/Simulink platform.Secondly,a 3-degrees-of-freedom(3-DOF)aircraft model is used as a maneuvering command generator,and the expanded elemental maneuver library is designed,so that the aircraft state reachable set can be obtained.Then,the game matrix is composed with the air combat situation evaluation function calculated according to the angle and range threats.Finally,a key point is that the objective function to be optimized is designed using the game mixed strategy,and the optimal mixed strategy is obtained by TLPIO.Significantly,the proposed TLPIO does not initialize the population randomly,but adopts the transfer learning method based on Kullback-Leibler(KL)divergence to initialize the population,which improves the search accuracy of the optimization algorithm.Besides,the convergence and time complexity of TLPIO are discussed.Comparison analysis with other classical optimization algorithms highlights the advantage of TLPIO.In the simulation of air combat,three initial scenarios are set,namely,opposite,offensive and defensive conditions.The effectiveness performance of the proposed autonomous maneuver decision method is verified by simulation results.
基金Supporting Project number(PNURSP2023R410)Princess Nourah bint Abdulrahman University,Riyadh,Saudi Arabia.supported by MRC,UK(MC_PC_17171)+9 种基金Royal Society,UK(RP202G0230)BHF,UK(AA/18/3/34220)Hope Foundation for Cancer Research,UK(RM60G0680)GCRF,UK(P202PF11)Sino‐UK Industrial Fund,UK(RP202G0289)LIAS,UK(P202ED10,P202RE969)Data Science Enhancement Fund,UK(P202RE237)Fight for Sight,UK(24NN201)Sino‐UK Education Fund,UK(OP202006)BBSRC,UK(RM32G0178B8).The funding of this work was provided by Princess Nourah bint Abdulrahman University Researchers Supporting Project number(PNURSP2023R410),Princess Nourah bint Abdulrahman University,Riyadh,Saudi Arabia.
文摘Currently,the improvement in AI is mainly related to deep learning techniques that are employed for the classification,identification,and quantification of patterns in clinical images.The deep learning models show more remarkable performance than the traditional methods for medical image processing tasks,such as skin cancer,colorectal cancer,brain tumour,cardiac disease,Breast cancer(BrC),and a few more.The manual diagnosis of medical issues always requires an expert and is also expensive.Therefore,developing some computer diagnosis techniques based on deep learning is essential.Breast cancer is the most frequently diagnosed cancer in females with a rapidly growing percentage.It is estimated that patients with BrC will rise to 70%in the next 20 years.If diagnosed at a later stage,the survival rate of patients with BrC is shallow.Hence,early detection is essential,increasing the survival rate to 50%.A new framework for BrC classification is presented that utilises deep learning and feature optimization.The significant steps of the presented framework include(i)hybrid contrast enhancement of acquired images,(ii)data augmentation to facilitate better learning of the Convolutional Neural Network(CNN)model,(iii)a pre‐trained ResNet‐101 model is utilised and modified according to selected dataset classes,(iv)deep transfer learning based model training for feature extraction,(v)the fusion of features using the proposed highly corrected function‐controlled canonical correlation analysis approach,and(vi)optimal feature selection using the modified Satin Bowerbird Optimization controlled Newton Raphson algorithm that finally classified using 10 machine learning classifiers.The experiments of the proposed framework have been carried out using the most critical and publicly available dataset,such as CBISDDSM,and obtained the best accuracy of 94.5%along with improved computation time.The comparison depicts that the presented method surpasses the current state‐ofthe‐art approaches.
基金supported by the Deanship of Scientific Research,at Imam Abdulrahman Bin Faisal University.Grant Number:2019-416-ASCS.
文摘Lung cancer is among the most frequent cancers in the world,with over one million deaths per year.Classification is required for lung cancer diagnosis and therapy to be effective,accurate,and reliable.Gene expression microarrays have made it possible to find genetic biomarkers for cancer diagnosis and prediction in a high-throughput manner.Machine Learning(ML)has been widely used to diagnose and classify lung cancer where the performance of ML methods is evaluated to identify the appropriate technique.Identifying and selecting the gene expression patterns can help in lung cancer diagnoses and classification.Normally,microarrays include several genes and may cause confusion or false prediction.Therefore,the Arithmetic Optimization Algorithm(AOA)is used to identify the optimal gene subset to reduce the number of selected genes.Which can allow the classifiers to yield the best performance for lung cancer classification.In addition,we proposed a modified version of AOA which can work effectively on the high dimensional dataset.In the modified AOA,the features are ranked by their weights and are used to initialize the AOA population.The exploitation process of AOA is then enhanced by developing a local search algorithm based on two neighborhood strategies.Finally,the efficiency of the proposed methods was evaluated on gene expression datasets related to Lung cancer using stratified 4-fold cross-validation.The method’s efficacy in selecting the optimal gene subset is underscored by its ability to maintain feature proportions between 10%to 25%.Moreover,the approach significantly enhances lung cancer prediction accuracy.For instance,Lung_Harvard1 achieved an accuracy of 97.5%,Lung_Harvard2 and Lung_Michigan datasets both achieved 100%,Lung_Adenocarcinoma obtained an accuracy of 88.2%,and Lung_Ontario achieved an accuracy of 87.5%.In conclusion,the results indicate the potential promise of the proposed modified AOA approach in classifying microarray cancer data.
文摘For training the present Neural Network(NN)models,the standard technique is to utilize decaying Learning Rates(LR).While the majority of these techniques commence with a large LR,they will decay multiple times over time.Decaying has been proved to enhance generalization as well as optimization.Other parameters,such as the network’s size,the number of hidden layers,drop-outs to avoid overfitting,batch size,and so on,are solely based on heuristics.This work has proposed Adaptive Teaching Learning Based(ATLB)Heuristic to identify the optimal hyperparameters for diverse networks.Here we consider three architec-tures Recurrent Neural Networks(RNN),Long Short Term Memory(LSTM),Bidirectional Long Short Term Memory(BiLSTM)of Deep Neural Networks for classification.The evaluation of the proposed ATLB is done through the various learning rate schedulers Cyclical Learning Rate(CLR),Hyperbolic Tangent Decay(HTD),and Toggle between Hyperbolic Tangent Decay and Triangular mode with Restarts(T-HTR)techniques.Experimental results have shown the performance improvement on the 20Newsgroup,Reuters Newswire and IMDB dataset.
基金Projects(61173122,61262032) supported by the National Natural Science Foundation of ChinaProjects(11JJ3067,12JJ2038) supported by the Natural Science Foundation of Hunan Province,China
文摘Low-rank matrix recovery is an important problem extensively studied in machine learning, data mining and computer vision communities. A novel method is proposed for low-rank matrix recovery, targeting at higher recovery accuracy and stronger theoretical guarantee. Specifically, the proposed method is based on a nonconvex optimization model, by solving the low-rank matrix which can be recovered from the noisy observation. To solve the model, an effective algorithm is derived by minimizing over the variables alternately. It is proved theoretically that this algorithm has stronger theoretical guarantee than the existing work. In natural image denoising experiments, the proposed method achieves lower recovery error than the two compared methods. The proposed low-rank matrix recovery method is also applied to solve two real-world problems, i.e., removing noise from verification code and removing watermark from images, in which the images recovered by the proposed method are less noisy than those of the two compared methods.
文摘In Internet of Things (IoT), large amount of data are processed andcommunicated through different network technologies. Wireless Body Area Networks (WBAN) plays pivotal role in the health care domain with an integration ofIoT and Artificial Intelligence (AI). The amalgamation of above mentioned toolshas taken the new peak in terms of diagnosis and treatment process especially inthe pandemic period. But the real challenges such as low latency, energy consumption high throughput still remains in the dark side of the research. This paperproposes a novel optimized cognitive learning based BAN model based on FogIoT technology as a real-time health monitoring systems with the increased network-life time. Energy and latency aware features of BAN have been extractedand used to train the proposed fog based learning algorithm to achieve low energyconsumption and low-latency scheduling algorithm. To test the proposed network,Fog-IoT-BAN test bed has been developed with the battery driven MICOTTboards interfaced with the health care sensors using Micro Python programming.The extensive experimentation is carried out using the above test beds and variousparameters such as accuracy, precision, recall, F1score and specificity has beencalculated along with QoS (quality of service) parameters such as latency, energyand throughput. To prove the superiority of the proposed framework, the performance of the proposed learning based framework has been compared with theother state-of-art classical learning frameworks and other existing Fog-BAN networks such as WORN, DARE, L-No-DEAF networks. Results proves the proposed framework has outperformed the other classical learning models in termsof accuracy and high False Alarm Rate (FAR), energy efficiency and latency.
基金Taif University Researchers Supporting Project Number(TURSP-2020/260),Taif University,Taif,Saudi Arabia.
文摘Millimeter wave communication works in the 30–300 GHz frequency range,and can obtain a very high bandwidth,which greatly improves the transmission rate of the communication system and becomes one of the key technologies of fifth-generation(5G).The smaller wavelength of the millimeter wave makes it possible to assemble a large number of antennas in a small aperture.The resulting array gain can compensate for the path loss of the millimeter wave.Utilizing this feature,the millimeter wave massive multiple-input multiple-output(MIMO)system uses a large antenna array at the base station.It enables the transmission of multiple data streams,making the system have a higher data transmission rate.In the millimeter wave massive MIMO system,the precoding technology uses the state information of the channel to adjust the transmission strategy at the transmitting end,and the receiving end performs equalization,so that users can better obtain the antenna multiplexing gain and improve the system capacity.This paper proposes an efficient algorithm based on machine learning(ML)for effective system performance in mmwave massive MIMO systems.The main idea is to optimize the adaptive connection structure to maximize the received signal power of each user and correlate the RF chain and base station antenna.Simulation results show that,the proposed algorithm effectively improved the system performance in terms of spectral efficiency and complexity as compared with existing algorithms.
文摘A chest radiology scan can significantly aid the early diagnosis and management of COVID-19 since the virus attacks the lungs.Chest X-ray(CXR)gained much interest after the COVID-19 outbreak thanks to its rapid imaging time,widespread availability,low cost,and portability.In radiological investigations,computer-aided diagnostic tools are implemented to reduce intra-and inter-observer variability.Using lately industrialized Artificial Intelligence(AI)algorithms and radiological techniques to diagnose and classify disease is advantageous.The current study develops an automatic identification and classification model for CXR pictures using Gaussian Fil-tering based Optimized Synergic Deep Learning using Remora Optimization Algorithm(GF-OSDL-ROA).This method is inclusive of preprocessing and classification based on optimization.The data is preprocessed using Gaussian filtering(GF)to remove any extraneous noise from the image’s edges.Then,the OSDL model is applied to classify the CXRs under different severity levels based on CXR data.The learning rate of OSDL is optimized with the help of ROA for COVID-19 diagnosis showing the novelty of the work.OSDL model,applied in this study,was validated using the COVID-19 dataset.The experiments were conducted upon the proposed OSDL model,which achieved a classification accuracy of 99.83%,while the current Convolutional Neural Network achieved less classification accuracy,i.e.,98.14%.
文摘The exponential growth of Internet of Things(IoT)and 5G networks has resulted in maximum users,and the role of cognitive radio has become pivotal in handling the crowded users.In this scenario,cognitive radio techniques such as spectrum sensing,spectrum sharing and dynamic spectrum access will become essential components in Wireless IoT communication.IoT devices must learn adaptively to the environment and extract the spectrum knowledge and inferred spectrum knowledge by appropriately changing communication parameters such as modulation index,frequency bands,coding rate etc.,to accommodate the above characteristics.Implementing the above learning methods on the embedded chip leads to high latency,high power consumption and more chip area utilisation.To overcome the problems mentioned above,we present DEEP HOLE Radio sys-tems,the intelligent system enabling the spectrum knowledge extraction from the unprocessed samples by the optimized deep learning models directly from the Radio Frequency(RF)environment.DEEP HOLE Radio provides(i)an opti-mized deep learning framework with a good trade-off between latency,power and utilization.(ii)Complete Hardware-Software architecture where the SoC’s coupled with radio transceivers for maximum performance.The experimentation has been carried out using GNURADIO software interfaced with Zynq-7000 devices mounting on ESP8266 radio transceivers with inbuilt Omni direc-tional antennas.The whole spectrum of knowledge has been extracted using GNU radio.These extracted features are used to train the proposed optimized deep learning models,which run parallel on Zynq-SoC 7000,consuming less area,power,latency and less utilization area.The proposed framework has been evaluated and compared with the existing frameworks such as RFLearn,Long Term Short Memory(LSTM),Convolutional Neural Networks(CNN)and Deep Neural Networks(DNN).The outcome shows that the proposed framework has outperformed the existing framework regarding the area,power and time.More-over,the experimental results show that the proposed framework decreases the delay,power and area by 15%,20%25%concerning the existing RFlearn and other hardware constraint frameworks.
文摘Surface coating is a critical procedure in the case of maintenance engineering. Ceramic coating of the wear areas is of the best practice which substantially enhances the Mean Time between Failure (MTBF). EN24 is a commercial grade alloy which is used for various industrial applications like sleeves, nuts, bolts, shafts, etc. EN24 is having comparatively low corrosion resistance, and ceramic coating of the wear and corroding areas of such parts is a best followed practice which highly improves the frequent failures. The coating quality mainly depends on the coating thickness, surface roughness and coating hardness which finally decides the operability. This paper describes an experimental investigation to effectively optimize the Atmospheric Plasma Spray process input parameters of Al<sub>2</sub>O<sub>3</sub>-40% TiO<sub>2</sub> coatings to get the best quality of coating on EN24 alloy steel substrate. The experiments are conducted with an Orthogonal Array (OA) design of experiments (DoE). In the current experiment, critical input parameters are considered and some of the vital output parameters are monitored accordingly and separate mathematical models are generated using regression analysis. The Analytic Hierarchy Process (AHP) method is used to generate weights for the individual objective functions and based on that, a combined objective function is made. An advanced optimization method, Teaching-Learning-Based Optimization algorithm (TLBO), is practically utilized to the combined objective function to optimize the values of input parameters to get the best output parameters. Confirmation tests are also conducted and their output results are compared with predicted values obtained through mathematical models. The dominating effects of Al<sub>2</sub>O<sub>3</sub>-40% TiO<sub>2</sub> spray parameters on output parameters: surface roughness, coating thickness and coating hardness are discussed in detail. It is concluded that the input parameters variation directly affects the characteristics of output parameters and any number of input as well as output parameters can be easily optimized using the current approach.
文摘SS304 is a commercial grade stainless steel which is used for various engineering applications like shafts, guides, jigs, fixtures, etc. Ceramic coating of the wear areas of such parts is a regular practice which significantly enhances the Mean Time Between Failure (MTBF). The final coating quality depends mainly on the coating thickness, surface roughness and hardness which ultimately decides the life. This paper presents an experimental study to effectively optimize the Atmospheric Plasma Spray (APS) process input parameters of Al<sub>2</sub>O<sub>3</sub>-40% TiO2 ceramic coatings to get the best quality of coating on commercial SS304 substrate. The experiments are conducted with a three-level L<sub>18</sub> Orthogonal Array (OA) Design of Experiments (DoE). Critical input parameters considered are: spray nozzle distance, substrate rotating speed, current of the arc, carrier gas flow and coating powder flow rate. The surface roughness, coating thickness and hardness are considered as the output parameters. Mathematical models are generated using regression analysis for individual output parameters. The Analytic Hierarchy Process (AHP) method is applied to generate weights for the individual objective functions and a combined objective function is generated. An advanced optimization method, Teaching-Learning-Based Optimization algorithm (TLBO), is applied to the combined objective function to optimize the values of input parameters to get the best output parameters and confirmation tests are conducted based on that. The significant effects of spray parameters on surface roughness, coating thickness and coating hardness are studied in detail.
文摘Learning to optimize(L2O)stands at the intersection of traditional optimization and machine learning,utilizing the capabilities of machine learning to enhance conventional optimization techniques.As real-world optimization problems frequently share common structures,L2O provides a tool to exploit these structures for better or faster solutions.This tutorial dives deep into L2O techniques,introducing how to accelerate optimization algorithms,promptly estimate the solutions,or even reshape the optimization problem itself,making it more adaptive to real-world applications.By considering the prerequisites for successful applications of L2O and the structure of the optimization problems at hand,this tutorial provides a comprehensive guide for practitioners and researchers alike.
基金Princess Nourah bint Abdulrahman University Researchers Supporting Project number(PNURSP2024R 343)PrincessNourah bint Abdulrahman University,Riyadh,Saudi ArabiaDeanship of Scientific Research at Northern Border University,Arar,Kingdom of Saudi Arabia,for funding this researchwork through the project number“NBU-FFR-2024-1092-02”.
文摘Phishing attacks present a persistent and evolving threat in the cybersecurity land-scape,necessitating the development of more sophisticated detection methods.Traditional machine learning approaches to phishing detection have relied heavily on feature engineering and have often fallen short in adapting to the dynamically changing patterns of phishingUniformResource Locator(URLs).Addressing these challenge,we introduce a framework that integrates the sequential data processing strengths of a Recurrent Neural Network(RNN)with the hyperparameter optimization prowess of theWhale Optimization Algorithm(WOA).Ourmodel capitalizes on an extensive Kaggle dataset,featuring over 11,000 URLs,each delineated by 30 attributes.The WOA’s hyperparameter optimization enhances the RNN’s performance,evidenced by a meticulous validation process.The results,encapsulated in precision,recall,and F1-score metrics,surpass baseline models,achieving an overall accuracy of 92%.This study not only demonstrates the RNN’s proficiency in learning complex patterns but also underscores the WOA’s effectiveness in refining machine learning models for the critical task of phishing detection.
基金partially supported by NSFC under Grant Nos.61832001 and 62272008ZTE Industry-University-Institute Fund Project。
文摘The query processing in distributed database management systems(DBMS)faces more challenges,such as more operators,and more factors in cost models and meta-data,than that in a single-node DMBS,in which query optimization is already an NP-hard problem.Learned query optimizers(mainly in the single-node DBMS)receive attention due to its capability to capture data distributions and flexible ways to avoid hard-craft rules in refinement and adaptation to new hardware.In this paper,we focus on extensions of learned query optimizers to distributed DBMSs.Specifically,we propose one possible but general architecture of the learned query optimizer in the distributed context and highlight differences from the learned optimizer in the single-node ones.In addition,we discuss the challenges and possible solutions.
文摘Purpose: This study aimed to enhance the prediction of container dwell time, a crucial factor for optimizing port operations, resource allocation, and supply chain efficiency. Determining an optimal learning rate for training Artificial Neural Networks (ANNs) has remained a challenging task due to the diverse sizes, complexity, and types of data involved. Design/Method/Approach: This research used a RandomizedSearchCV algorithm, a random search approach, to bridge this knowledge gap. The algorithm was applied to container dwell time data from the TOS system of the Port of Tema, which included 307,594 container records from 2014 to 2022. Findings: The RandomizedSearchCV method outperformed standard training methods both in terms of reducing training time and improving prediction accuracy, highlighting the significant role of the constant learning rate as a hyperparameter. Research Limitations and Implications: Although the study provides promising outcomes, the results are limited to the data extracted from the Port of Tema and may differ in other contexts. Further research is needed to generalize these findings across various port systems. Originality/Value: This research underscores the potential of RandomizedSearchCV as a valuable tool for optimizing ANN training in container dwell time prediction. It also accentuates the significance of automated learning rate selection, offering novel insights into the optimization of container dwell time prediction, with implications for improving port efficiency and supply chain operations.
基金partially supported by the Science and Technology Innovation 2030-Key Project of“New Generation Artificial Intelligence”(Grant No.2018AAA0102403)the National Natural Science Foundation of China(Grant Nos.U20B2071,91948204,T2121003,and U1913602)。
文摘For unmanned aerial vehicle(UAV)swarm dynamic combat,swarm antagonistic motion control and attack target allocation are extremely challenging sub-tasks.In this paper,the competitive learning pigeon-inspired optimization(CLPIO)algorithm is proposed to handle the cooperative dynamic combat problem,which integrates the distributed swarm antagonistic motion and centralized attack target allocation.Moreover,the threshold trigger strategy is presented to switch two sub-tasks.To seek a feasible and optimal combat scheme,a dynamic game approach combined with hawk grouping mechanism and situation assessment between sub-groups is designed to guide the solution of the optimal attack scheme,and the model of swarm antagonistic motion imitating pigeon’s intelligence is proposed to form a confrontation situation.The analysis of the CLPIO algorithm shows its convergence in theory and the comparison with the other four metaheuristic algorithms shows its superiority in solving the mixed Nash equilibrium problem.Finally,numerical simulation verifis that the proposed methods can provide an effective combat scheme in the set scenario.
基金Project supported by the Science and Technology Innovation 2030-Key Project of“New Generation Artificial Intelligence,”China(No.018AAA0102303)the National Natural Science Foundation of China(Nos.91948204,91648205,U1913602,and U19B2033)the Aeronautical Foundation of China(No.20185851022)。
文摘We propose multi-objective social learning pigeon-inspired optimization(MSLPIO)and apply it to obstacle avoidance for unmanned aerial vehicle(UAV)formation.In the algorithm,each pigeon learns from the better pigeon but not necessarily the global best one in the update process.A social learning factor is added to the map and compass operator and the landmark operator.In addition,a dimension-dependent parameter setting method is adopted to improve the blindness of parameter setting.We simulate the flight process of five UAVs in a complex obstacle environment.Results verify the effectiveness of the proposed method.MSLPIO has better convergence performance compared with the improved multi-objective pigeon-inspired optimization and the improved non-dominated sorting genetic algorithm.
基金supported by National Natural Science Foundation of China(No.51377103)the technology project of State Grid Corporation of China:Research on Multi-Level Decomposition Coordination of the Pareto Set of Multi-Objective Optimization Problem in Bulk Power System(No.SGSXDKYDWKJ2015-001)the support from State Energy Smart Grid R&D Center(SHANGHAI)
文摘This paper employs an efficacious analytical tool,adaptive simplified human learning optimization(ASHLO)algorithm,to solve optimal power flow(OPF)problem in AC/DC hybrid power system,considering valve-point loading effects of generators,carbon tax,and prohibited operating zones of generators,respectively.ASHLO algorithm,involves random learning operator,individual learning operator,social learning operator and adaptive strategies.To compare and analyze the computation performance of the ASHLO method,the proposed ASHLO method and other heuristic intelligent optimization methods are employed to solve OPF problem on the modified IEEE 30-bus and 118-bus AC/DC hybrid test system.Numerical results indicate that the ASHLO method has good convergent property and robustness.Meanwhile,the impacts of wind speeds and locations of HVDC transmission line integrated into the AC network on the OPF results are systematically analyzed.
基金Projects(61573144,61773165,61673175,61174040)supported by the National Natural Science Foundation of ChinaProject(222201717006)supported by the Fundamental Research Funds for the Central Universities,China
文摘The hybrid flow shop scheduling problem with unrelated parallel machine is a typical NP-hard combinatorial optimization problem, and it exists widely in chemical, manufacturing and pharmaceutical industry. In this work, a novel mathematic model for the hybrid flow shop scheduling problem with unrelated parallel machine(HFSPUPM) was proposed. Additionally, an effective hybrid estimation of distribution algorithm was proposed to solve the HFSPUPM, taking advantage of the features in the mathematic model. In the optimization algorithm, a new individual representation method was adopted. The(EDA) structure was used for global search while the teaching learning based optimization(TLBO) strategy was used for local search. Based on the structure of the HFSPUPM, this work presents a series of discrete operations. Simulation results show the effectiveness of the proposed hybrid algorithm compared with other algorithms.
基金supported by the National Natural Science Foundation of China(No.:52177028)Aeronautical Science Foundation of China(No.201907051002)+1 种基金the Fundamental Research Funds for the Central Universities,China(No.YWF21BJJ522)the Major Program of the National Natural Science Foundation of China(No.51890882).
文摘To maximize the power density of the electric propulsion motor in aerospace application,this paper proposes a novel Dynamic Neighborhood Genetic Learning Particle Swarm Optimization(DNGL-PSO)for the motor design,which can deal with the insufficient population diversity and non-global optimal solution issues.The DNGL-PSO framework is composed of the dynamic neighborhood module and the particle update module.To improve the population diversity,the dynamic neighborhood strategy is first proposed,which combines the local neighborhood exemplar generation mechanism and the shuffling mechanism.The local neighborhood exemplar generation mechanism enlarges the search range of the algorithm in the solution space,thus obtaining highquality exemplars.Meanwhile,when the global optimal solution cannot update its fitness value,the shuffling mechanism module is triggered to dynamically change the local neighborhood members.The roulette wheel selection operator is introduced into the shuffling mechanism to ensure that particles with larger fitness value are selected with a higher probability and remain in the local neighborhood.Then,the global learning based particle update approach is proposed,which can achieve a good balance between the expansion of the search range in the early stage and the acceleration of local convergence in the later stage.Finally,the optimization design of the electric propulsion motor is conducted to verify the effectiveness of the proposed DNGL-PSO.The simulation results show that the proposed DNGL-PSO has excellent adaptability,optimization efficiency and global optimization capability,while the optimized electric propulsion motor has a high power density of 5.207 kW/kg with the efficiency of 96.12%.