The phenomenon of a target echo peak overlapping with the backscattered echo peak significantly undermines the detection range and precision of underwater laser fuzes.To overcome this issue,we propose a four-quadrant ...The phenomenon of a target echo peak overlapping with the backscattered echo peak significantly undermines the detection range and precision of underwater laser fuzes.To overcome this issue,we propose a four-quadrant dual-beam circumferential scanning laser fuze to distinguish various interference signals and provide more real-time data for the backscatter filtering algorithm.This enhances the algorithm loading capability of the fuze.In order to address the problem of insufficient filtering capacity in existing linear backscatter filtering algorithms,we develop a nonlinear backscattering adaptive filter based on the spline adaptive filter least mean square(SAF-LMS)algorithm.We also designed an algorithm pause module to retain the original trend of the target echo peak,improving the time discrimination accuracy and anti-interference capability of the fuze.Finally,experiments are conducted with varying signal-to-noise ratios of the original underwater target echo signals.The experimental results show that the average signal-to-noise ratio before and after filtering can be improved by more than31 d B,with an increase of up to 76%in extreme detection distance.展开更多
A dandelion algorithm(DA) is a recently developed intelligent optimization algorithm for function optimization problems. Many of its parameters need to be set by experience in DA,which might not be appropriate for all...A dandelion algorithm(DA) is a recently developed intelligent optimization algorithm for function optimization problems. Many of its parameters need to be set by experience in DA,which might not be appropriate for all optimization problems. A self-adapting and efficient dandelion algorithm is proposed in this work to lower the number of DA's parameters and simplify DA's structure. Only the normal sowing operator is retained;while the other operators are discarded. An adaptive seeding radius strategy is designed for the core dandelion. The results show that the proposed algorithm achieves better performance on the standard test functions with less time consumption than its competitive peers. In addition, the proposed algorithm is applied to feature selection for credit card fraud detection(CCFD), and the results indicate that it can obtain higher classification and detection performance than the-state-of-the-art methods.展开更多
Cloud computing is a dynamic and rapidly evolving field,where the demand for resources fluctuates continuously.This paper delves into the imperative need for adaptability in the allocation of resources to applications...Cloud computing is a dynamic and rapidly evolving field,where the demand for resources fluctuates continuously.This paper delves into the imperative need for adaptability in the allocation of resources to applications and services within cloud computing environments.The motivation stems from the pressing issue of accommodating fluctuating levels of user demand efficiently.By adhering to the proposed resource allocation method,we aim to achieve a substantial reduction in energy consumption.This reduction hinges on the precise and efficient allocation of resources to the tasks that require those most,aligning with the broader goal of sustainable and eco-friendly cloud computing systems.To enhance the resource allocation process,we introduce a novel knowledge-based optimization algorithm.In this study,we rigorously evaluate its efficacy by comparing it to existing algorithms,including the Flower Pollination Algorithm(FPA),Spark Lion Whale Optimization(SLWO),and Firefly Algo-rithm.Our findings reveal that our proposed algorithm,Knowledge Based Flower Pollination Algorithm(KB-FPA),consistently outperforms these conventional methods in both resource allocation efficiency and energy consumption reduction.This paper underscores the profound significance of resource allocation in the realm of cloud computing.By addressing the critical issue of adaptability and energy efficiency,it lays the groundwork for a more sustainable future in cloud computing systems.Our contribution to the field lies in the introduction of a new resource allocation strategy,offering the potential for significantly improved efficiency and sustainability within cloud computing infrastructures.展开更多
In view of the complex marine environment of navigation,especially in the case of multiple static and dynamic obstacles,the traditional obstacle avoidance algorithms applied to unmanned surface vehicles(USV)are prone ...In view of the complex marine environment of navigation,especially in the case of multiple static and dynamic obstacles,the traditional obstacle avoidance algorithms applied to unmanned surface vehicles(USV)are prone to fall into the trap of local optimization.Therefore,this paper proposes an improved artificial potential field(APF)algorithm,which uses 5G communication technology to communicate between the USV and the control center.The algorithm introduces the USV discrimination mechanism to avoid the USV falling into local optimization when the USV encounter different obstacles in different scenarios.Considering the various scenarios between the USV and other dynamic obstacles such as vessels in the process of performing tasks,the algorithm introduces the concept of dynamic artificial potential field.For the multiple obstacles encountered in the process of USV sailing,based on the International Regulations for Preventing Collisions at Sea(COLREGS),the USV determines whether the next step will fall into local optimization through the discriminationmechanism.The local potential field of the USV will dynamically adjust,and the reverse virtual gravitational potential field will be added to prevent it from falling into the local optimization and avoid collisions.The objective function and cost function are designed at the same time,so that the USV can smoothly switch between the global path and the local obstacle avoidance.The simulation results show that the improved APF algorithm proposed in this paper can successfully avoid various obstacles in the complex marine environment,and take navigation time and economic cost into account.展开更多
In Wireless Sensor Networks(WSNs),Clustering process is widely utilized for increasing the lifespan with sustained energy stability during data transmission.Several clustering protocols were devised for extending netw...In Wireless Sensor Networks(WSNs),Clustering process is widely utilized for increasing the lifespan with sustained energy stability during data transmission.Several clustering protocols were devised for extending network lifetime,but most of them failed in handling the problem of fixed clustering,static rounds,and inadequate Cluster Head(CH)selection criteria which consumes more energy.In this paper,Stochastic Ranking Improved Teaching-Learning and Adaptive Grasshopper Optimization Algorithm(SRITL-AGOA)-based Clustering Scheme for energy stabilization and extending network lifespan.This SRITL-AGOA selected CH depending on the weightage of factors such as node mobility degree,neighbour's density distance to sink,single-hop or multihop communication and Residual Energy(RE)that directly influences the energy consumption of sensor nodes.In specific,Grasshopper Optimization Algorithm(GOA)is improved through tangent-based nonlinear strategy for enhancing the ability of global optimization.On the other hand,stochastic ranking and violation constraint handling strategies are embedded into Teaching-Learning-based Optimization Algorithm(TLOA)for improving its exploitation tendencies.Then,SR and VCH improved TLOA is embedded into the exploitation phase of AGOA for selecting better CH by maintaining better balance amid exploration and exploitation.Simulation results confirmed that the proposed SRITL-AGOA improved throughput by 21.86%,network stability by 18.94%,load balancing by 16.14%with minimized energy depletion by19.21%,compared to the competitive CH selection approaches.展开更多
Frequency modulated continuous wave(FMCW)radar is an advantageous sensor scheme for target estimation and environmental perception.However,existing algorithms based on discrete Fourier transform(DFT),multiple signal c...Frequency modulated continuous wave(FMCW)radar is an advantageous sensor scheme for target estimation and environmental perception.However,existing algorithms based on discrete Fourier transform(DFT),multiple signal classification(MUSIC)and compressed sensing,etc.,cannot achieve both low complexity and high resolution simultaneously.This paper proposes an efficient 2-D MUSIC algorithm for super-resolution target estimation/tracking based on FMCW radar.Firstly,we enhance the efficiency of 2-D MUSIC azimuth-range spectrum estimation by incorporating 2-D DFT and multi-level resolution searching strategy.Secondly,we apply the gradient descent method to tightly integrate the spatial continuity of object motion into spectrum estimation when processing multi-epoch radar data,which improves the efficiency of continuous target tracking.These two approaches have improved the algorithm efficiency by nearly 2-4 orders of magnitude without losing accuracy and resolution.Simulation experiments are conducted to validate the effectiveness of the algorithm in both single-epoch estimation and multi-epoch tracking scenarios.展开更多
For unachievable tracking problems, where the system output cannot precisely track a given reference, achieving the best possible approximation for the reference trajectory becomes the objective. This study aims to in...For unachievable tracking problems, where the system output cannot precisely track a given reference, achieving the best possible approximation for the reference trajectory becomes the objective. This study aims to investigate solutions using the Ptype learning control scheme. Initially, we demonstrate the necessity of gradient information for achieving the best approximation.Subsequently, we propose an input-output-driven learning gain design to handle the imprecise gradients of a class of uncertain systems. However, it is discovered that the desired performance may not be attainable when faced with incomplete information.To address this issue, an extended iterative learning control scheme is introduced. In this scheme, the tracking errors are modified through output data sampling, which incorporates lowmemory footprints and offers flexibility in learning gain design.The input sequence is shown to converge towards the desired input, resulting in an output that is closest to the given reference in the least square sense. Numerical simulations are provided to validate the theoretical findings.展开更多
The development of technologies such as big data and blockchain has brought convenience to life,but at the same time,privacy and security issues are becoming more and more prominent.The K-anonymity algorithm is an eff...The development of technologies such as big data and blockchain has brought convenience to life,but at the same time,privacy and security issues are becoming more and more prominent.The K-anonymity algorithm is an effective and low computational complexity privacy-preserving algorithm that can safeguard users’privacy by anonymizing big data.However,the algorithm currently suffers from the problem of focusing only on improving user privacy while ignoring data availability.In addition,ignoring the impact of quasi-identified attributes on sensitive attributes causes the usability of the processed data on statistical analysis to be reduced.Based on this,we propose a new K-anonymity algorithm to solve the privacy security problem in the context of big data,while guaranteeing improved data usability.Specifically,we construct a new information loss function based on the information quantity theory.Considering that different quasi-identification attributes have different impacts on sensitive attributes,we set weights for each quasi-identification attribute when designing the information loss function.In addition,to reduce information loss,we improve K-anonymity in two ways.First,we make the loss of information smaller than in the original table while guaranteeing privacy based on common artificial intelligence algorithms,i.e.,greedy algorithm and 2-means clustering algorithm.In addition,we improve the 2-means clustering algorithm by designing a mean-center method to select the initial center of mass.Meanwhile,we design the K-anonymity algorithm of this scheme based on the constructed information loss function,the improved 2-means clustering algorithm,and the greedy algorithm,which reduces the information loss.Finally,we experimentally demonstrate the effectiveness of the algorithm in improving the effect of 2-means clustering and reducing information loss.展开更多
In blood or bone marrow,leukemia is a form of cancer.A person with leukemia has an expansion of white blood cells(WBCs).It primarily affects children and rarely affects adults.Treatment depends on the type of leukemia...In blood or bone marrow,leukemia is a form of cancer.A person with leukemia has an expansion of white blood cells(WBCs).It primarily affects children and rarely affects adults.Treatment depends on the type of leukemia and the extent to which cancer has established throughout the body.Identifying leukemia in the initial stage is vital to providing timely patient care.Medical image-analysis-related approaches grant safer,quicker,and less costly solutions while ignoring the difficulties of these invasive processes.It can be simple to generalize Computer vision(CV)-based and image-processing techniques and eradicate human error.Many researchers have implemented computer-aided diagnosticmethods andmachine learning(ML)for laboratory image analysis,hopefully overcoming the limitations of late leukemia detection and determining its subgroups.This study establishes a Marine Predators Algorithm with Deep Learning Leukemia Cancer Classification(MPADL-LCC)algorithm onMedical Images.The projectedMPADL-LCC system uses a bilateral filtering(BF)technique to pre-process medical images.The MPADL-LCC system uses Faster SqueezeNet withMarine Predators Algorithm(MPA)as a hyperparameter optimizer for feature extraction.Lastly,the denoising autoencoder(DAE)methodology can be executed to accurately detect and classify leukemia cancer.The hyperparameter tuning process using MPA helps enhance leukemia cancer classification performance.Simulation results are compared with other recent approaches concerning various measurements and the MPADL-LCC algorithm exhibits the best results over other recent approaches.展开更多
Lung cancer is among the most frequent cancers in the world,with over one million deaths per year.Classification is required for lung cancer diagnosis and therapy to be effective,accurate,and reliable.Gene expression ...Lung cancer is among the most frequent cancers in the world,with over one million deaths per year.Classification is required for lung cancer diagnosis and therapy to be effective,accurate,and reliable.Gene expression microarrays have made it possible to find genetic biomarkers for cancer diagnosis and prediction in a high-throughput manner.Machine Learning(ML)has been widely used to diagnose and classify lung cancer where the performance of ML methods is evaluated to identify the appropriate technique.Identifying and selecting the gene expression patterns can help in lung cancer diagnoses and classification.Normally,microarrays include several genes and may cause confusion or false prediction.Therefore,the Arithmetic Optimization Algorithm(AOA)is used to identify the optimal gene subset to reduce the number of selected genes.Which can allow the classifiers to yield the best performance for lung cancer classification.In addition,we proposed a modified version of AOA which can work effectively on the high dimensional dataset.In the modified AOA,the features are ranked by their weights and are used to initialize the AOA population.The exploitation process of AOA is then enhanced by developing a local search algorithm based on two neighborhood strategies.Finally,the efficiency of the proposed methods was evaluated on gene expression datasets related to Lung cancer using stratified 4-fold cross-validation.The method’s efficacy in selecting the optimal gene subset is underscored by its ability to maintain feature proportions between 10%to 25%.Moreover,the approach significantly enhances lung cancer prediction accuracy.For instance,Lung_Harvard1 achieved an accuracy of 97.5%,Lung_Harvard2 and Lung_Michigan datasets both achieved 100%,Lung_Adenocarcinoma obtained an accuracy of 88.2%,and Lung_Ontario achieved an accuracy of 87.5%.In conclusion,the results indicate the potential promise of the proposed modified AOA approach in classifying microarray cancer data.展开更多
Current methodologies for cleaning wind power anomaly data exhibit limited capabilities in identifying abnormal data within extensive datasets and struggle to accommodate the considerable variability and intricacy of ...Current methodologies for cleaning wind power anomaly data exhibit limited capabilities in identifying abnormal data within extensive datasets and struggle to accommodate the considerable variability and intricacy of wind farm data.Consequently,a method for cleaning wind power anomaly data by combining image processing with community detection algorithms(CWPAD-IPCDA)is proposed.To precisely identify and initially clean anomalous data,wind power curve(WPC)images are converted into graph structures,which employ the Louvain community recognition algorithm and graph-theoretic methods for community detection and segmentation.Furthermore,the mathematical morphology operation(MMO)determines the main part of the initially cleaned wind power curve images and maps them back to the normal wind power points to complete the final cleaning.The CWPAD-IPCDA method was applied to clean datasets from 25 wind turbines(WTs)in two wind farms in northwest China to validate its feasibility.A comparison was conducted using density-based spatial clustering of applications with noise(DBSCAN)algorithm,an improved isolation forest algorithm,and an image-based(IB)algorithm.The experimental results demonstrate that the CWPAD-IPCDA method surpasses the other three algorithms,achieving an approximately 7.23%higher average data cleaning rate.The mean value of the sum of the squared errors(SSE)of the dataset after cleaning is approximately 6.887 lower than that of the other algorithms.Moreover,the mean of overall accuracy,as measured by the F1-score,exceeds that of the other methods by approximately 10.49%;this indicates that the CWPAD-IPCDA method is more conducive to improving the accuracy and reliability of wind power curve modeling and wind farm power forecasting.展开更多
The chirp sub-bottom profiler,for its high resolution,easy accessibility and cost-effectiveness,has been widely used in acoustic detection.In this paper,the acoustic impedance and grain size compositions were obtained...The chirp sub-bottom profiler,for its high resolution,easy accessibility and cost-effectiveness,has been widely used in acoustic detection.In this paper,the acoustic impedance and grain size compositions were obtained based on the chirp sub-bottom profiler data collected in the Chukchi Plateau area during the 11th Arctic Expedition of China.The time-domain adaptive search matching algorithm was used and validated on our established theoretical model.The misfit between the inversion result and the theoretical model is less than 0.067%.The grain size was calculated according to the empirical relationship between the acoustic impedance and the grain size of the sediment.The average acoustic impedance of sub-seafloor strata is 2.5026×10^(6) kg(s m^(2))^(-1)and the average grain size(θvalue)of the seafloor surface sediment is 7.1498,indicating the predominant occurrence of very fine silt sediment in the study area.Comparison of the inversion results and the laboratory measurements of nearby borehole samples shows that they are in general agreement.展开更多
Active Magnetic Bearing(AMB) is a kind of electromagnetic support that makes the rotor movement frictionless and can suppress rotor vibration by controlling the magnetic force. The most common approach to restrain the...Active Magnetic Bearing(AMB) is a kind of electromagnetic support that makes the rotor movement frictionless and can suppress rotor vibration by controlling the magnetic force. The most common approach to restrain the rotor vibration in AMBs is to adopt a notch filter or adaptive filter in the AMB controller. However, these methods cannot obtain the precise amplitude and phase of the compensation current. Thus, they are not so effective in terms of suppressing the vibrations of the fundamental and other harmonic orders over the whole speed range. To improve the vibration suppression performance of AMBs,an adaptive filter based on Least Mean Square(LMS) is applied to extract the vibration signals from the rotor displacement signal. An Iterative Search Algorithm(ISA) is proposed in this paper to obtain the corresponding relationship between the compensation current and vibration signals. The ISA is responsible for searching the compensating amplitude and shifting phase online for the LMS filter, enabling the AMB controller to generate the corresponding compensation force for vibration suppression. The results of ISA are recorded to suppress vibration using the Look-Up Table(LUT) in variable speed range. Comprehensive simulations and experimental validations are carried out in fixed and variable speed range, and the results demonstrate that by employing the ISA, vibrations of the fundamental and other harmonic orders are suppressed effectively.展开更多
It is still a huge challenge for traditional Pareto-dominatedmany-objective optimization algorithms to solve manyobjective optimization problems because these algorithms hardly maintain the balance between convergence...It is still a huge challenge for traditional Pareto-dominatedmany-objective optimization algorithms to solve manyobjective optimization problems because these algorithms hardly maintain the balance between convergence and diversity and can only find a group of solutions focused on a small area on the Pareto front,resulting in poor performance of those algorithms.For this reason,we propose a reference vector-assisted algorithmwith an adaptive niche dominance relation,for short MaOEA-AR.The new dominance relation forms a niche based on the angle between candidate solutions.By comparing these solutions,the solutionwith the best convergence is found to be the non-dominated solution to improve the selection pressure.In reproduction,a mutation strategy of k-bit crossover and hybrid mutation is used to generate high-quality offspring.On 23 test problems with up to 15-objective,we compared the proposed algorithm with five state-of-the-art algorithms.The experimental results verified that the proposed algorithm is competitive.展开更多
The exploration of urban underground spaces is of great significance to urban planning,geological disaster prevention,resource exploration and environmental monitoring.However,due to the existing of severe interferenc...The exploration of urban underground spaces is of great significance to urban planning,geological disaster prevention,resource exploration and environmental monitoring.However,due to the existing of severe interferences,conventional seismic methods cannot adapt to the complex urban environment well.Since adopting the single-node data acquisition method and taking the seismic ambient noise as the signal,the microtremor horizontal-to-vertical spectral ratio(HVSR)method can effectively avoid the strong interference problems caused by the complex urban environment,which could obtain information such as S-wave velocity and thickness of underground formations by fitting the microtremor HVSR curve.Nevertheless,HVSR curve inversion is a multi-parameter curve fitting process.And conventional inversion methods can easily converge to the local minimum,which will directly affect the reliability of the inversion results.Thus,the authors propose a HVSR inversion method based on the multimodal forest optimization algorithm,which uses the efficient clustering technique and locates the global optimum quickly.Tests on synthetic data show that the inversion results of the proposed method are consistent with the forward model.Both the adaption and stability to the abnormal layer velocity model are demonstrated.The results of the real field data are also verified by the drilling information.展开更多
This article mainly investigates the fuzzy optimization robust control issue for nonlinear networked systems characterized by the interval type-2(IT2)fuzzy technique under a differential evolution algorithm.To provide...This article mainly investigates the fuzzy optimization robust control issue for nonlinear networked systems characterized by the interval type-2(IT2)fuzzy technique under a differential evolution algorithm.To provide a more reasonable utilization of the constrained communication channel,a novel adaptive memory event-triggered(AMET)mechanism is developed,where two event-triggered thresholds can be dynamically adjusted in the light of the current system information and the transmitted historical data.Sufficient conditions with less conservative design of the fuzzy imperfect premise matching(IPM)controller are presented by introducing the Wirtinger-based integral inequality,the information of membership functions(MFs)and slack matrices.Subsequently,under the IPM policy,a new MFs intelligent optimization technique that takes advantage of the differential evolution algorithm is first provided for IT2 TakagiSugeno(T-S)fuzzy systems to update the fuzzy controller MFs in real-time and achieve a better system control effect.Finally,simulation results demonstrate that the proposed control scheme can obtain better system performance in the case of using fewer communication resources.展开更多
In the last five years,there has been a V-shaped recovery in the number of papers on congestion control algorithms on the Internet.In this paper,congestion problems on the Internet are discussed,such as congestion col...In the last five years,there has been a V-shaped recovery in the number of papers on congestion control algorithms on the Internet.In this paper,congestion problems on the Internet are discussed,such as congestion collapse and bufferbloat from the perspective of the necessity of congestion control algorithms.The typical congestion control algorithms are introduced,and the research areas and methods of congestion control algorithms are described.Recent research trends and future prospects of congestion control algorithms are also presented.展开更多
Compositional data, such as relative information, is a crucial aspect of machine learning and other related fields. It is typically recorded as closed data or sums to a constant, like 100%. The statistical linear mode...Compositional data, such as relative information, is a crucial aspect of machine learning and other related fields. It is typically recorded as closed data or sums to a constant, like 100%. The statistical linear model is the most used technique for identifying hidden relationships between underlying random variables of interest. However, data quality is a significant challenge in machine learning, especially when missing data is present. The linear regression model is a commonly used statistical modeling technique used in various applications to find relationships between variables of interest. When estimating linear regression parameters which are useful for things like future prediction and partial effects analysis of independent variables, maximum likelihood estimation (MLE) is the method of choice. However, many datasets contain missing observations, which can lead to costly and time-consuming data recovery. To address this issue, the expectation-maximization (EM) algorithm has been suggested as a solution for situations including missing data. The EM algorithm repeatedly finds the best estimates of parameters in statistical models that depend on variables or data that have not been observed. This is called maximum likelihood or maximum a posteriori (MAP). Using the present estimate as input, the expectation (E) step constructs a log-likelihood function. Finding the parameters that maximize the anticipated log-likelihood, as determined in the E step, is the job of the maximization (M) phase. This study looked at how well the EM algorithm worked on a made-up compositional dataset with missing observations. It used both the robust least square version and ordinary least square regression techniques. The efficacy of the EM algorithm was compared with two alternative imputation techniques, k-Nearest Neighbor (k-NN) and mean imputation (), in terms of Aitchison distances and covariance.展开更多
To counter the defect of traditional genetic algorithms, an improved adaptivegenetic algorithm with the criterion of premature convergence is provided. The occurrence ofpremature convergence is forecasted using colony...To counter the defect of traditional genetic algorithms, an improved adaptivegenetic algorithm with the criterion of premature convergence is provided. The occurrence ofpremature convergence is forecasted using colony entropy and colony variance. When prematureconvergence occurs, new individuals are generated in proper scale randomly based on superiorindividuals in the colony. We use these new individuals to replace some individuals in the oldcolony. The updated individuals account for 30 percent - 40 percent of all individuals and the sizeof scale is related to the distribution of the extreme value of the target function. Simulationtests show that there is much improvement in the speed of convergence and the probability of globalconvergence.展开更多
基金supported by the 2021 Open Project Fund of Science and Technology on Electromechanical Dynamic Control Laboratory,grant number 212-C-J-F-QT-2022-0020China Postdoctoral Science Foundation,grant number 2021M701713+1 种基金Postgraduate Research&Practice Innovation Program of Jiangsu Province,grant number KYCX23_0511the Jiangsu Funding Program for Excellent Postdoctoral Talent,grant number 20220ZB245。
文摘The phenomenon of a target echo peak overlapping with the backscattered echo peak significantly undermines the detection range and precision of underwater laser fuzes.To overcome this issue,we propose a four-quadrant dual-beam circumferential scanning laser fuze to distinguish various interference signals and provide more real-time data for the backscatter filtering algorithm.This enhances the algorithm loading capability of the fuze.In order to address the problem of insufficient filtering capacity in existing linear backscatter filtering algorithms,we develop a nonlinear backscattering adaptive filter based on the spline adaptive filter least mean square(SAF-LMS)algorithm.We also designed an algorithm pause module to retain the original trend of the target echo peak,improving the time discrimination accuracy and anti-interference capability of the fuze.Finally,experiments are conducted with varying signal-to-noise ratios of the original underwater target echo signals.The experimental results show that the average signal-to-noise ratio before and after filtering can be improved by more than31 d B,with an increase of up to 76%in extreme detection distance.
基金supported by the Institutional Fund Projects(IFPIP-1481-611-1443)the Key Projects of Natural Science Research in Anhui Higher Education Institutions(2022AH051909)+1 种基金the Provincial Quality Project of Colleges and Universities in Anhui Province(2022sdxx020,2022xqhz044)Bengbu University 2021 High-Level Scientific Research and Cultivation Project(2021pyxm04)。
文摘A dandelion algorithm(DA) is a recently developed intelligent optimization algorithm for function optimization problems. Many of its parameters need to be set by experience in DA,which might not be appropriate for all optimization problems. A self-adapting and efficient dandelion algorithm is proposed in this work to lower the number of DA's parameters and simplify DA's structure. Only the normal sowing operator is retained;while the other operators are discarded. An adaptive seeding radius strategy is designed for the core dandelion. The results show that the proposed algorithm achieves better performance on the standard test functions with less time consumption than its competitive peers. In addition, the proposed algorithm is applied to feature selection for credit card fraud detection(CCFD), and the results indicate that it can obtain higher classification and detection performance than the-state-of-the-art methods.
基金supported by the Ministerio Espanol de Ciencia e Innovación under Project Number PID2020-115570GB-C22 MCIN/AEI/10.13039/501100011033 and by the Cátedra de Empresa Tecnología para las Personas(UGR-Fujitsu).
文摘Cloud computing is a dynamic and rapidly evolving field,where the demand for resources fluctuates continuously.This paper delves into the imperative need for adaptability in the allocation of resources to applications and services within cloud computing environments.The motivation stems from the pressing issue of accommodating fluctuating levels of user demand efficiently.By adhering to the proposed resource allocation method,we aim to achieve a substantial reduction in energy consumption.This reduction hinges on the precise and efficient allocation of resources to the tasks that require those most,aligning with the broader goal of sustainable and eco-friendly cloud computing systems.To enhance the resource allocation process,we introduce a novel knowledge-based optimization algorithm.In this study,we rigorously evaluate its efficacy by comparing it to existing algorithms,including the Flower Pollination Algorithm(FPA),Spark Lion Whale Optimization(SLWO),and Firefly Algo-rithm.Our findings reveal that our proposed algorithm,Knowledge Based Flower Pollination Algorithm(KB-FPA),consistently outperforms these conventional methods in both resource allocation efficiency and energy consumption reduction.This paper underscores the profound significance of resource allocation in the realm of cloud computing.By addressing the critical issue of adaptability and energy efficiency,it lays the groundwork for a more sustainable future in cloud computing systems.Our contribution to the field lies in the introduction of a new resource allocation strategy,offering the potential for significantly improved efficiency and sustainability within cloud computing infrastructures.
基金This work was supported by the Postdoctoral Fund of FDCT,Macao(Grant No.0003/2021/APD).Any opinions,findings and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect those of the sponsor.
文摘In view of the complex marine environment of navigation,especially in the case of multiple static and dynamic obstacles,the traditional obstacle avoidance algorithms applied to unmanned surface vehicles(USV)are prone to fall into the trap of local optimization.Therefore,this paper proposes an improved artificial potential field(APF)algorithm,which uses 5G communication technology to communicate between the USV and the control center.The algorithm introduces the USV discrimination mechanism to avoid the USV falling into local optimization when the USV encounter different obstacles in different scenarios.Considering the various scenarios between the USV and other dynamic obstacles such as vessels in the process of performing tasks,the algorithm introduces the concept of dynamic artificial potential field.For the multiple obstacles encountered in the process of USV sailing,based on the International Regulations for Preventing Collisions at Sea(COLREGS),the USV determines whether the next step will fall into local optimization through the discriminationmechanism.The local potential field of the USV will dynamically adjust,and the reverse virtual gravitational potential field will be added to prevent it from falling into the local optimization and avoid collisions.The objective function and cost function are designed at the same time,so that the USV can smoothly switch between the global path and the local obstacle avoidance.The simulation results show that the improved APF algorithm proposed in this paper can successfully avoid various obstacles in the complex marine environment,and take navigation time and economic cost into account.
文摘In Wireless Sensor Networks(WSNs),Clustering process is widely utilized for increasing the lifespan with sustained energy stability during data transmission.Several clustering protocols were devised for extending network lifetime,but most of them failed in handling the problem of fixed clustering,static rounds,and inadequate Cluster Head(CH)selection criteria which consumes more energy.In this paper,Stochastic Ranking Improved Teaching-Learning and Adaptive Grasshopper Optimization Algorithm(SRITL-AGOA)-based Clustering Scheme for energy stabilization and extending network lifespan.This SRITL-AGOA selected CH depending on the weightage of factors such as node mobility degree,neighbour's density distance to sink,single-hop or multihop communication and Residual Energy(RE)that directly influences the energy consumption of sensor nodes.In specific,Grasshopper Optimization Algorithm(GOA)is improved through tangent-based nonlinear strategy for enhancing the ability of global optimization.On the other hand,stochastic ranking and violation constraint handling strategies are embedded into Teaching-Learning-based Optimization Algorithm(TLOA)for improving its exploitation tendencies.Then,SR and VCH improved TLOA is embedded into the exploitation phase of AGOA for selecting better CH by maintaining better balance amid exploration and exploitation.Simulation results confirmed that the proposed SRITL-AGOA improved throughput by 21.86%,network stability by 18.94%,load balancing by 16.14%with minimized energy depletion by19.21%,compared to the competitive CH selection approaches.
基金funded by the National Natural Science Foundation of China,grant number 42074176,U1939204。
文摘Frequency modulated continuous wave(FMCW)radar is an advantageous sensor scheme for target estimation and environmental perception.However,existing algorithms based on discrete Fourier transform(DFT),multiple signal classification(MUSIC)and compressed sensing,etc.,cannot achieve both low complexity and high resolution simultaneously.This paper proposes an efficient 2-D MUSIC algorithm for super-resolution target estimation/tracking based on FMCW radar.Firstly,we enhance the efficiency of 2-D MUSIC azimuth-range spectrum estimation by incorporating 2-D DFT and multi-level resolution searching strategy.Secondly,we apply the gradient descent method to tightly integrate the spatial continuity of object motion into spectrum estimation when processing multi-epoch radar data,which improves the efficiency of continuous target tracking.These two approaches have improved the algorithm efficiency by nearly 2-4 orders of magnitude without losing accuracy and resolution.Simulation experiments are conducted to validate the effectiveness of the algorithm in both single-epoch estimation and multi-epoch tracking scenarios.
基金supported by the National Natural Science Foundation of China (62173333, 12271522)Beijing Natural Science Foundation (Z210002)the Research Fund of Renmin University of China (2021030187)。
文摘For unachievable tracking problems, where the system output cannot precisely track a given reference, achieving the best possible approximation for the reference trajectory becomes the objective. This study aims to investigate solutions using the Ptype learning control scheme. Initially, we demonstrate the necessity of gradient information for achieving the best approximation.Subsequently, we propose an input-output-driven learning gain design to handle the imprecise gradients of a class of uncertain systems. However, it is discovered that the desired performance may not be attainable when faced with incomplete information.To address this issue, an extended iterative learning control scheme is introduced. In this scheme, the tracking errors are modified through output data sampling, which incorporates lowmemory footprints and offers flexibility in learning gain design.The input sequence is shown to converge towards the desired input, resulting in an output that is closest to the given reference in the least square sense. Numerical simulations are provided to validate the theoretical findings.
基金Foundation of National Natural Science Foundation of China(62202118)Scientific and Technological Research Projects from Guizhou Education Department([2023]003)+1 种基金Guizhou Provincial Department of Science and Technology Hundred Levels of Innovative Talents Project(GCC[2023]018)Top Technology Talent Project from Guizhou Education Department([2022]073).
文摘The development of technologies such as big data and blockchain has brought convenience to life,but at the same time,privacy and security issues are becoming more and more prominent.The K-anonymity algorithm is an effective and low computational complexity privacy-preserving algorithm that can safeguard users’privacy by anonymizing big data.However,the algorithm currently suffers from the problem of focusing only on improving user privacy while ignoring data availability.In addition,ignoring the impact of quasi-identified attributes on sensitive attributes causes the usability of the processed data on statistical analysis to be reduced.Based on this,we propose a new K-anonymity algorithm to solve the privacy security problem in the context of big data,while guaranteeing improved data usability.Specifically,we construct a new information loss function based on the information quantity theory.Considering that different quasi-identification attributes have different impacts on sensitive attributes,we set weights for each quasi-identification attribute when designing the information loss function.In addition,to reduce information loss,we improve K-anonymity in two ways.First,we make the loss of information smaller than in the original table while guaranteeing privacy based on common artificial intelligence algorithms,i.e.,greedy algorithm and 2-means clustering algorithm.In addition,we improve the 2-means clustering algorithm by designing a mean-center method to select the initial center of mass.Meanwhile,we design the K-anonymity algorithm of this scheme based on the constructed information loss function,the improved 2-means clustering algorithm,and the greedy algorithm,which reduces the information loss.Finally,we experimentally demonstrate the effectiveness of the algorithm in improving the effect of 2-means clustering and reducing information loss.
基金funded by Researchers Supporting Program at King Saud University,(RSPD2024R809).
文摘In blood or bone marrow,leukemia is a form of cancer.A person with leukemia has an expansion of white blood cells(WBCs).It primarily affects children and rarely affects adults.Treatment depends on the type of leukemia and the extent to which cancer has established throughout the body.Identifying leukemia in the initial stage is vital to providing timely patient care.Medical image-analysis-related approaches grant safer,quicker,and less costly solutions while ignoring the difficulties of these invasive processes.It can be simple to generalize Computer vision(CV)-based and image-processing techniques and eradicate human error.Many researchers have implemented computer-aided diagnosticmethods andmachine learning(ML)for laboratory image analysis,hopefully overcoming the limitations of late leukemia detection and determining its subgroups.This study establishes a Marine Predators Algorithm with Deep Learning Leukemia Cancer Classification(MPADL-LCC)algorithm onMedical Images.The projectedMPADL-LCC system uses a bilateral filtering(BF)technique to pre-process medical images.The MPADL-LCC system uses Faster SqueezeNet withMarine Predators Algorithm(MPA)as a hyperparameter optimizer for feature extraction.Lastly,the denoising autoencoder(DAE)methodology can be executed to accurately detect and classify leukemia cancer.The hyperparameter tuning process using MPA helps enhance leukemia cancer classification performance.Simulation results are compared with other recent approaches concerning various measurements and the MPADL-LCC algorithm exhibits the best results over other recent approaches.
基金supported by the Deanship of Scientific Research,at Imam Abdulrahman Bin Faisal University.Grant Number:2019-416-ASCS.
文摘Lung cancer is among the most frequent cancers in the world,with over one million deaths per year.Classification is required for lung cancer diagnosis and therapy to be effective,accurate,and reliable.Gene expression microarrays have made it possible to find genetic biomarkers for cancer diagnosis and prediction in a high-throughput manner.Machine Learning(ML)has been widely used to diagnose and classify lung cancer where the performance of ML methods is evaluated to identify the appropriate technique.Identifying and selecting the gene expression patterns can help in lung cancer diagnoses and classification.Normally,microarrays include several genes and may cause confusion or false prediction.Therefore,the Arithmetic Optimization Algorithm(AOA)is used to identify the optimal gene subset to reduce the number of selected genes.Which can allow the classifiers to yield the best performance for lung cancer classification.In addition,we proposed a modified version of AOA which can work effectively on the high dimensional dataset.In the modified AOA,the features are ranked by their weights and are used to initialize the AOA population.The exploitation process of AOA is then enhanced by developing a local search algorithm based on two neighborhood strategies.Finally,the efficiency of the proposed methods was evaluated on gene expression datasets related to Lung cancer using stratified 4-fold cross-validation.The method’s efficacy in selecting the optimal gene subset is underscored by its ability to maintain feature proportions between 10%to 25%.Moreover,the approach significantly enhances lung cancer prediction accuracy.For instance,Lung_Harvard1 achieved an accuracy of 97.5%,Lung_Harvard2 and Lung_Michigan datasets both achieved 100%,Lung_Adenocarcinoma obtained an accuracy of 88.2%,and Lung_Ontario achieved an accuracy of 87.5%.In conclusion,the results indicate the potential promise of the proposed modified AOA approach in classifying microarray cancer data.
基金supported by the National Natural Science Foundation of China(Project No.51767018)Natural Science Foundation of Gansu Province(Project No.23JRRA836).
文摘Current methodologies for cleaning wind power anomaly data exhibit limited capabilities in identifying abnormal data within extensive datasets and struggle to accommodate the considerable variability and intricacy of wind farm data.Consequently,a method for cleaning wind power anomaly data by combining image processing with community detection algorithms(CWPAD-IPCDA)is proposed.To precisely identify and initially clean anomalous data,wind power curve(WPC)images are converted into graph structures,which employ the Louvain community recognition algorithm and graph-theoretic methods for community detection and segmentation.Furthermore,the mathematical morphology operation(MMO)determines the main part of the initially cleaned wind power curve images and maps them back to the normal wind power points to complete the final cleaning.The CWPAD-IPCDA method was applied to clean datasets from 25 wind turbines(WTs)in two wind farms in northwest China to validate its feasibility.A comparison was conducted using density-based spatial clustering of applications with noise(DBSCAN)algorithm,an improved isolation forest algorithm,and an image-based(IB)algorithm.The experimental results demonstrate that the CWPAD-IPCDA method surpasses the other three algorithms,achieving an approximately 7.23%higher average data cleaning rate.The mean value of the sum of the squared errors(SSE)of the dataset after cleaning is approximately 6.887 lower than that of the other algorithms.Moreover,the mean of overall accuracy,as measured by the F1-score,exceeds that of the other methods by approximately 10.49%;this indicates that the CWPAD-IPCDA method is more conducive to improving the accuracy and reliability of wind power curve modeling and wind farm power forecasting.
基金supported by the National Key R&D Program of China (No.2021YFC2801202)the National Natural Science Foundation of China (No.42076224)the Fundamental Research Funds for the Central Universities (No.202262012)。
文摘The chirp sub-bottom profiler,for its high resolution,easy accessibility and cost-effectiveness,has been widely used in acoustic detection.In this paper,the acoustic impedance and grain size compositions were obtained based on the chirp sub-bottom profiler data collected in the Chukchi Plateau area during the 11th Arctic Expedition of China.The time-domain adaptive search matching algorithm was used and validated on our established theoretical model.The misfit between the inversion result and the theoretical model is less than 0.067%.The grain size was calculated according to the empirical relationship between the acoustic impedance and the grain size of the sediment.The average acoustic impedance of sub-seafloor strata is 2.5026×10^(6) kg(s m^(2))^(-1)and the average grain size(θvalue)of the seafloor surface sediment is 7.1498,indicating the predominant occurrence of very fine silt sediment in the study area.Comparison of the inversion results and the laboratory measurements of nearby borehole samples shows that they are in general agreement.
基金supported by the Natural Science Foundation of China (U22A20214)。
文摘Active Magnetic Bearing(AMB) is a kind of electromagnetic support that makes the rotor movement frictionless and can suppress rotor vibration by controlling the magnetic force. The most common approach to restrain the rotor vibration in AMBs is to adopt a notch filter or adaptive filter in the AMB controller. However, these methods cannot obtain the precise amplitude and phase of the compensation current. Thus, they are not so effective in terms of suppressing the vibrations of the fundamental and other harmonic orders over the whole speed range. To improve the vibration suppression performance of AMBs,an adaptive filter based on Least Mean Square(LMS) is applied to extract the vibration signals from the rotor displacement signal. An Iterative Search Algorithm(ISA) is proposed in this paper to obtain the corresponding relationship between the compensation current and vibration signals. The ISA is responsible for searching the compensating amplitude and shifting phase online for the LMS filter, enabling the AMB controller to generate the corresponding compensation force for vibration suppression. The results of ISA are recorded to suppress vibration using the Look-Up Table(LUT) in variable speed range. Comprehensive simulations and experimental validations are carried out in fixed and variable speed range, and the results demonstrate that by employing the ISA, vibrations of the fundamental and other harmonic orders are suppressed effectively.
基金supported by the National Natural Science Foundation of China(Grant No.61976101)the University Natural Science Research Project of Anhui Province(Grant No.2023AH040056)+4 种基金the Natural Science Research Project of Anhui Province(Graduate Research Project,Grant No.YJS20210463)the Funding Plan for Scientic Research Activities of Academic and Technical Leaders and Reserve Candidates in Anhui Province(Grant No.2021H264)the Top Talent Project of Disciplines(Majors)in Colleges and Universities in Anhui Province(Grant No.gxbjZD2022021)the University Synergy Innovation Program of Anhui Province,China(GXXT-2022-033)supported by the Innovation Fund for Postgraduates of Huaibei Normal University(Grant Nos.cx2022041,yx2021023,CX2023043).
文摘It is still a huge challenge for traditional Pareto-dominatedmany-objective optimization algorithms to solve manyobjective optimization problems because these algorithms hardly maintain the balance between convergence and diversity and can only find a group of solutions focused on a small area on the Pareto front,resulting in poor performance of those algorithms.For this reason,we propose a reference vector-assisted algorithmwith an adaptive niche dominance relation,for short MaOEA-AR.The new dominance relation forms a niche based on the angle between candidate solutions.By comparing these solutions,the solutionwith the best convergence is found to be the non-dominated solution to improve the selection pressure.In reproduction,a mutation strategy of k-bit crossover and hybrid mutation is used to generate high-quality offspring.On 23 test problems with up to 15-objective,we compared the proposed algorithm with five state-of-the-art algorithms.The experimental results verified that the proposed algorithm is competitive.
基金Supported by projects of National Natural Science Foundation of China(No.42074150)National Key Research and Development Program of China(No.2023YFC3707901)Futian District Integrated Ground Collapse Monitoring and Early Warning System Construction Project(No.FTCG2023000209).
文摘The exploration of urban underground spaces is of great significance to urban planning,geological disaster prevention,resource exploration and environmental monitoring.However,due to the existing of severe interferences,conventional seismic methods cannot adapt to the complex urban environment well.Since adopting the single-node data acquisition method and taking the seismic ambient noise as the signal,the microtremor horizontal-to-vertical spectral ratio(HVSR)method can effectively avoid the strong interference problems caused by the complex urban environment,which could obtain information such as S-wave velocity and thickness of underground formations by fitting the microtremor HVSR curve.Nevertheless,HVSR curve inversion is a multi-parameter curve fitting process.And conventional inversion methods can easily converge to the local minimum,which will directly affect the reliability of the inversion results.Thus,the authors propose a HVSR inversion method based on the multimodal forest optimization algorithm,which uses the efficient clustering technique and locates the global optimum quickly.Tests on synthetic data show that the inversion results of the proposed method are consistent with the forward model.Both the adaption and stability to the abnormal layer velocity model are demonstrated.The results of the real field data are also verified by the drilling information.
基金supported by the National Natural Science Foundation of China(61973105,62373137)。
文摘This article mainly investigates the fuzzy optimization robust control issue for nonlinear networked systems characterized by the interval type-2(IT2)fuzzy technique under a differential evolution algorithm.To provide a more reasonable utilization of the constrained communication channel,a novel adaptive memory event-triggered(AMET)mechanism is developed,where two event-triggered thresholds can be dynamically adjusted in the light of the current system information and the transmitted historical data.Sufficient conditions with less conservative design of the fuzzy imperfect premise matching(IPM)controller are presented by introducing the Wirtinger-based integral inequality,the information of membership functions(MFs)and slack matrices.Subsequently,under the IPM policy,a new MFs intelligent optimization technique that takes advantage of the differential evolution algorithm is first provided for IT2 TakagiSugeno(T-S)fuzzy systems to update the fuzzy controller MFs in real-time and achieve a better system control effect.Finally,simulation results demonstrate that the proposed control scheme can obtain better system performance in the case of using fewer communication resources.
基金supported by JSPS Grants-in-Aid for Scientific Research JP20K11786 and JP21KK0202.
文摘In the last five years,there has been a V-shaped recovery in the number of papers on congestion control algorithms on the Internet.In this paper,congestion problems on the Internet are discussed,such as congestion collapse and bufferbloat from the perspective of the necessity of congestion control algorithms.The typical congestion control algorithms are introduced,and the research areas and methods of congestion control algorithms are described.Recent research trends and future prospects of congestion control algorithms are also presented.
文摘Compositional data, such as relative information, is a crucial aspect of machine learning and other related fields. It is typically recorded as closed data or sums to a constant, like 100%. The statistical linear model is the most used technique for identifying hidden relationships between underlying random variables of interest. However, data quality is a significant challenge in machine learning, especially when missing data is present. The linear regression model is a commonly used statistical modeling technique used in various applications to find relationships between variables of interest. When estimating linear regression parameters which are useful for things like future prediction and partial effects analysis of independent variables, maximum likelihood estimation (MLE) is the method of choice. However, many datasets contain missing observations, which can lead to costly and time-consuming data recovery. To address this issue, the expectation-maximization (EM) algorithm has been suggested as a solution for situations including missing data. The EM algorithm repeatedly finds the best estimates of parameters in statistical models that depend on variables or data that have not been observed. This is called maximum likelihood or maximum a posteriori (MAP). Using the present estimate as input, the expectation (E) step constructs a log-likelihood function. Finding the parameters that maximize the anticipated log-likelihood, as determined in the E step, is the job of the maximization (M) phase. This study looked at how well the EM algorithm worked on a made-up compositional dataset with missing observations. It used both the robust least square version and ordinary least square regression techniques. The efficacy of the EM algorithm was compared with two alternative imputation techniques, k-Nearest Neighbor (k-NN) and mean imputation (), in terms of Aitchison distances and covariance.
基金The Natural Science Foundation of Jiangsu Province (BK99011).
文摘To counter the defect of traditional genetic algorithms, an improved adaptivegenetic algorithm with the criterion of premature convergence is provided. The occurrence ofpremature convergence is forecasted using colony entropy and colony variance. When prematureconvergence occurs, new individuals are generated in proper scale randomly based on superiorindividuals in the colony. We use these new individuals to replace some individuals in the oldcolony. The updated individuals account for 30 percent - 40 percent of all individuals and the sizeof scale is related to the distribution of the extreme value of the target function. Simulationtests show that there is much improvement in the speed of convergence and the probability of globalconvergence.