High-dimensional datasets present significant challenges for classification tasks.Dimensionality reduction,a crucial aspect of data preprocessing,has gained substantial attention due to its ability to improve classifi...High-dimensional datasets present significant challenges for classification tasks.Dimensionality reduction,a crucial aspect of data preprocessing,has gained substantial attention due to its ability to improve classification per-formance.However,identifying the optimal features within high-dimensional datasets remains a computationally demanding task,necessitating the use of efficient algorithms.This paper introduces the Arithmetic Optimization Algorithm(AOA),a novel approach for finding the optimal feature subset.AOA is specifically modified to address feature selection problems based on a transfer function.Additionally,two enhancements are incorporated into the AOA algorithm to overcome limitations such as limited precision,slow convergence,and susceptibility to local optima.The first enhancement proposes a new method for selecting solutions to be improved during the search process.This method effectively improves the original algorithm’s accuracy and convergence speed.The second enhancement introduces a local search with neighborhood strategies(AOA_NBH)during the AOA exploitation phase.AOA_NBH explores the vast search space,aiding the algorithm in escaping local optima.Our results demonstrate that incorporating neighborhood methods enhances the output and achieves significant improvement over state-of-the-art methods.展开更多
Various feature selection algorithms are usually employed to improve classification models’overall performance.Optimization algorithms typically accompany such algorithms to select the optimal set of features.Among t...Various feature selection algorithms are usually employed to improve classification models’overall performance.Optimization algorithms typically accompany such algorithms to select the optimal set of features.Among the most currently attractive trends within optimization algorithms are hybrid metaheuristics.The present paper presents two Stages of Local Search models for feature selection based on WOA(Whale Optimization Algorithm)and Great Deluge(GD).GD Algorithm is integrated with the WOA algorithm to improve exploitation by identifying the most promising regions during the search.Another version is employed using the best solution found by the WOA algorithm and exploited by the GD algorithm.In addition,disruptive selection(DS)is employed to select the solutions from the population for local search.DS is chosen to maintain the diversity of the population via enhancing low and high-quality solutions.Fifteen(15)standard benchmark datasets provided by the University of California Irvine(UCI)repository were used in evaluating the proposed approaches’performance.Next,a comparison was made with four population-based algorithms as wrapper feature selection methods from the literature.The proposed techniques have proved their efficiency in enhancing classification accuracy compared to other wrapper methods.Hence,the WOA can search effectively in the feature space and choose the most relevant attributes for classification tasks.展开更多
文摘High-dimensional datasets present significant challenges for classification tasks.Dimensionality reduction,a crucial aspect of data preprocessing,has gained substantial attention due to its ability to improve classification per-formance.However,identifying the optimal features within high-dimensional datasets remains a computationally demanding task,necessitating the use of efficient algorithms.This paper introduces the Arithmetic Optimization Algorithm(AOA),a novel approach for finding the optimal feature subset.AOA is specifically modified to address feature selection problems based on a transfer function.Additionally,two enhancements are incorporated into the AOA algorithm to overcome limitations such as limited precision,slow convergence,and susceptibility to local optima.The first enhancement proposes a new method for selecting solutions to be improved during the search process.This method effectively improves the original algorithm’s accuracy and convergence speed.The second enhancement introduces a local search with neighborhood strategies(AOA_NBH)during the AOA exploitation phase.AOA_NBH explores the vast search space,aiding the algorithm in escaping local optima.Our results demonstrate that incorporating neighborhood methods enhances the output and achieves significant improvement over state-of-the-art methods.
基金This research is part of a project funded by Imam Abdulrahman Bin Faisal University,under Grant Number 2020-083-BASRC.
文摘Various feature selection algorithms are usually employed to improve classification models’overall performance.Optimization algorithms typically accompany such algorithms to select the optimal set of features.Among the most currently attractive trends within optimization algorithms are hybrid metaheuristics.The present paper presents two Stages of Local Search models for feature selection based on WOA(Whale Optimization Algorithm)and Great Deluge(GD).GD Algorithm is integrated with the WOA algorithm to improve exploitation by identifying the most promising regions during the search.Another version is employed using the best solution found by the WOA algorithm and exploited by the GD algorithm.In addition,disruptive selection(DS)is employed to select the solutions from the population for local search.DS is chosen to maintain the diversity of the population via enhancing low and high-quality solutions.Fifteen(15)standard benchmark datasets provided by the University of California Irvine(UCI)repository were used in evaluating the proposed approaches’performance.Next,a comparison was made with four population-based algorithms as wrapper feature selection methods from the literature.The proposed techniques have proved their efficiency in enhancing classification accuracy compared to other wrapper methods.Hence,the WOA can search effectively in the feature space and choose the most relevant attributes for classification tasks.