Determination of Shear Bond strength(SBS)at interlayer of double-layer asphalt concrete is crucial in flexible pavement structures.The study used three Machine Learning(ML)models,including K-Nearest Neighbors(KNN),Ext...Determination of Shear Bond strength(SBS)at interlayer of double-layer asphalt concrete is crucial in flexible pavement structures.The study used three Machine Learning(ML)models,including K-Nearest Neighbors(KNN),Extra Trees(ET),and Light Gradient Boosting Machine(LGBM),to predict SBS based on easily determinable input parameters.Also,the Grid Search technique was employed for hyper-parameter tuning of the ML models,and cross-validation and learning curve analysis were used for training the models.The models were built on a database of 240 experimental results and three input variables:temperature,normal pressure,and tack coat rate.Model validation was performed using three statistical criteria:the coefficient of determination(R2),the Root Mean Square Error(RMSE),and the mean absolute error(MAE).Additionally,SHAP analysis was also used to validate the importance of the input variables in the prediction of the SBS.Results show that these models accurately predict SBS,with LGBM providing outstanding performance.SHAP(Shapley Additive explanation)analysis for LGBM indicates that temperature is the most influential factor on SBS.Consequently,the proposed ML models can quickly and accurately predict SBS between two layers of asphalt concrete,serving practical applications in flexible pavement structure design.展开更多
Traditional large-scale multi-objective optimization algorithms(LSMOEAs)encounter difficulties when dealing with sparse large-scale multi-objective optimization problems(SLM-OPs)where most decision variables are zero....Traditional large-scale multi-objective optimization algorithms(LSMOEAs)encounter difficulties when dealing with sparse large-scale multi-objective optimization problems(SLM-OPs)where most decision variables are zero.As a result,many algorithms use a two-layer encoding approach to optimize binary variable Mask and real variable Dec separately.Nevertheless,existing optimizers often focus on locating non-zero variable posi-tions to optimize the binary variables Mask.However,approxi-mating the sparse distribution of real Pareto optimal solutions does not necessarily mean that the objective function is optimized.In data mining,it is common to mine frequent itemsets appear-ing together in a dataset to reveal the correlation between data.Inspired by this,we propose a novel two-layer encoding learning swarm optimizer based on frequent itemsets(TELSO)to address these SLMOPs.TELSO mined the frequent terms of multiple particles with better target values to find mask combinations that can obtain better objective values for fast convergence.Experi-mental results on five real-world problems and eight benchmark sets demonstrate that TELSO outperforms existing state-of-the-art sparse large-scale multi-objective evolutionary algorithms(SLMOEAs)in terms of performance and convergence speed.展开更多
Passive detection of low-slow-small(LSS)targets is easily interfered by direct signal and multipath clutter,and the traditional clutter suppression method has the contradiction between step size and convergence rate.I...Passive detection of low-slow-small(LSS)targets is easily interfered by direct signal and multipath clutter,and the traditional clutter suppression method has the contradiction between step size and convergence rate.In this paper,a frequency domain clutter suppression algorithm based on sparse adaptive filtering is proposed.The pulse compression operation between the error signal and the input reference signal is added to the cost function as a sparsity constraint,and the criterion for filter weight updating is improved to obtain a purer echo signal.At the same time,the step size and penalty factor are brought into the adaptive iteration process,and the input data is used to drive the adaptive changes of parameters such as step size.The proposed algorithm has a small amount of calculation,which improves the robustness to parameters such as step size,reduces the weight error of the filter and has a good clutter suppression performance.展开更多
Most of the neural network architectures are based on human experience,which requires a long and tedious trial-and-error process.Neural architecture search(NAS)attempts to detect effective architectures without human ...Most of the neural network architectures are based on human experience,which requires a long and tedious trial-and-error process.Neural architecture search(NAS)attempts to detect effective architectures without human intervention.Evolutionary algorithms(EAs)for NAS can find better solutions than human-designed architectures by exploring a large search space for possible architectures.Using multiobjective EAs for NAS,optimal neural architectures that meet various performance criteria can be explored and discovered efficiently.Furthermore,hardware-accelerated NAS methods can improve the efficiency of the NAS.While existing reviews have mainly focused on different strategies to complete NAS,a few studies have explored the use of EAs for NAS.In this paper,we summarize and explore the use of EAs for NAS,as well as large-scale multiobjective optimization strategies and hardware-accelerated NAS methods.NAS performs well in healthcare applications,such as medical image analysis,classification of disease diagnosis,and health monitoring.EAs for NAS can automate the search process and optimize multiple objectives simultaneously in a given healthcare task.Deep neural network has been successfully used in healthcare,but it lacks interpretability.Medical data is highly sensitive,and privacy leaks are frequently reported in the healthcare industry.To solve these problems,in healthcare,we propose an interpretable neuroevolution framework based on federated learning to address search efficiency and privacy protection.Moreover,we also point out future research directions for evolutionary NAS.Overall,for researchers who want to use EAs to optimize NNs in healthcare,we analyze the advantages and disadvantages of doing so to provide detailed guidance,and propose an interpretable privacy-preserving framework for healthcare applications.展开更多
Improvement of integrated battlefield situational awareness in complex environments involving dynamic factors such as restricted communications and electromagnetic interference(EMI)has become a contentious research pr...Improvement of integrated battlefield situational awareness in complex environments involving dynamic factors such as restricted communications and electromagnetic interference(EMI)has become a contentious research problem.In certain mission environments,due to the impact of many interference sources on real-time communication or mission requirements such as the need to implement communication regulations,the mission stages are represented as a dynamic combination of several communication-available and communication-unavailable stages.Furthermore,the data interaction between unmanned aerial vehicles(UAVs)can only be performed in specific communication-available stages.Traditional cooperative search algorithms cannot handle such situations well.To solve this problem,this study constructed a distributed model predictive control(DMPC)architecture for a collaborative control of UAVs and used the Voronoi diagram generation method to re-plan the search areas of all UAVs in real time to avoid repetition of search areas and UAV collisions while improving the search efficiency and safety factor.An attention mechanism ant-colony optimization(AACO)algorithm is proposed for UAV search-control decision planning.The search strategy is adaptively updated by introducing an attention mechanism for regular instruction information,a priori information,and emergent information of the mission to satisfy different search expectations to the maximum extent.Simulation results show that the proposed algorithm achieves better search performance than traditional algorithms in restricted communication constraint scenarios.展开更多
Sparse large-scale multi-objective optimization problems(SLMOPs)are common in science and engineering.However,the large-scale problem represents the high dimensionality of the decision space,requiring algorithms to tr...Sparse large-scale multi-objective optimization problems(SLMOPs)are common in science and engineering.However,the large-scale problem represents the high dimensionality of the decision space,requiring algorithms to traverse vast expanse with limited computational resources.Furthermore,in the context of sparse,most variables in Pareto optimal solutions are zero,making it difficult for algorithms to identify non-zero variables efficiently.This paper is dedicated to addressing the challenges posed by SLMOPs.To start,we introduce innovative objective functions customized to mine maximum and minimum candidate sets.This substantial enhancement dramatically improves the efficacy of frequent pattern mining.In this way,selecting candidate sets is no longer based on the quantity of nonzero variables they contain but on a higher proportion of nonzero variables within specific dimensions.Additionally,we unveil a novel approach to association rule mining,which delves into the intricate relationships between non-zero variables.This novel methodology aids in identifying sparse distributions that can potentially expedite reductions in the objective function value.We extensively tested our algorithm across eight benchmark problems and four real-world SLMOPs.The results demonstrate that our approach achieves competitive solutions across various challenges.展开更多
Aiming at the practical application of Unmanned Underwater Vehicle(UUV)in underwater combat,this paper proposes a battlefield ambush scene with UUV considering ocean current.Firstly,by establishing these mathematical ...Aiming at the practical application of Unmanned Underwater Vehicle(UUV)in underwater combat,this paper proposes a battlefield ambush scene with UUV considering ocean current.Firstly,by establishing these mathematical models of ocean current environment,target movement,and sonar detection,the probability calculation methods of single UUV searching target and multiple UUV cooperatively searching target are given respectively.Then,based on the Hybrid Quantum-behaved Particle Swarm Optimization(HQPSO)algorithm,the path with the highest target search probability is found.Finally,through simulation calculations,the influence of different UUV parameters and target parameters on the target search probability is analyzed,and the minimum number of UUVs that need to be deployed to complete the ambush task is demonstrated,and the optimal search path scheme is obtained.The method proposed in this paper provides a theoretical basis for the practical application of UUV in the future combat.展开更多
In differentiable search architecture search methods,a more efficient search space design can significantly improve the performance of the searched architecture,thus requiring people to carefully define the search spa...In differentiable search architecture search methods,a more efficient search space design can significantly improve the performance of the searched architecture,thus requiring people to carefully define the search space with different complexity according to various operations.Meanwhile rationalizing the search strategies to explore the well-defined search space will further improve the speed and efficiency of architecture search.With this in mind,we propose a faster and more efficient differentiable architecture search method,AllegroNAS.Firstly,we introduce a more efficient search space enriched by the introduction of two redefined convolution modules.Secondly,we utilize a more efficient architectural parameter regularization method,mitigating the overfitting problem during the search process and reducing the error brought about by gradient approximation.Meanwhile,we introduce a natural exponential cosine annealing method to make the learning rate of the neural network training process more suitable for the search procedure.Moreover,group convolution and data augmentation are employed to reduce the computational cost.Finally,through extensive experiments on several public datasets,we demonstrate that our method can more swiftly search for better-performing neural network architectures in a more efficient search space,thus validating the effectiveness of our approach.展开更多
Contract Bridge,a four-player imperfect information game,comprises two phases:bidding and playing.While computer programs excel at playing,bidding presents a challenging aspect due to the need for information exchange...Contract Bridge,a four-player imperfect information game,comprises two phases:bidding and playing.While computer programs excel at playing,bidding presents a challenging aspect due to the need for information exchange with partners and interference with communication of opponents.In this work,we introduce a Bridge bidding agent that combines supervised learning,deep reinforcement learning via self-play,and a test-time search approach.Our experiments demonstrate that our agent outperforms WBridge5,a highly regarded computer Bridge software that has won multiple world championships,by a performance of 0.98 IMPs(international match points)per deal over 10000 deals,with a much cost-effective approach.The performance significantly surpasses previous state-of-the-art(0.85 IMPs per deal).Note 0.1 IMPs per deal is a significant improvement in Bridge bidding.展开更多
In this paper,we reconstruct strongly-decaying block sparse signals by the block generalized orthogonal matching pursuit(BgOMP)algorithm in the l2-bounded noise case.Under some restraints on the minimum magnitude of t...In this paper,we reconstruct strongly-decaying block sparse signals by the block generalized orthogonal matching pursuit(BgOMP)algorithm in the l2-bounded noise case.Under some restraints on the minimum magnitude of the nonzero elements of the strongly-decaying block sparse signal,if the sensing matrix satisfies the the block restricted isometry property(block-RIP),then arbitrary strongly-decaying block sparse signals can be accurately and steadily reconstructed by the BgOMP algorithm in iterations.Furthermore,we conjecture that this condition is sharp.展开更多
Although many multi-view clustering(MVC) algorithms with acceptable performances have been presented, to the best of our knowledge, nearly all of them need to be fed with the correct number of clusters. In addition, t...Although many multi-view clustering(MVC) algorithms with acceptable performances have been presented, to the best of our knowledge, nearly all of them need to be fed with the correct number of clusters. In addition, these existing algorithms create only the hard and fuzzy partitions for multi-view objects,which are often located in highly-overlapping areas of multi-view feature space. The adoption of hard and fuzzy partition ignores the ambiguity and uncertainty in the assignment of objects, likely leading to performance degradation. To address these issues, we propose a novel sparse reconstructive multi-view evidential clustering algorithm(SRMVEC). Based on a sparse reconstructive procedure, SRMVEC learns a shared affinity matrix across views, and maps multi-view objects to a 2-dimensional humanreadable chart by calculating 2 newly defined mathematical metrics for each object. From this chart, users can detect the number of clusters and select several objects existing in the dataset as cluster centers. Then, SRMVEC derives a credal partition under the framework of evidence theory, improving the fault tolerance of clustering. Ablation studies show the benefits of adopting the sparse reconstructive procedure and evidence theory. Besides,SRMVEC delivers effectiveness on benchmark datasets by outperforming some state-of-the-art methods.展开更多
Many methods based on deep learning have achieved amazing results in image sentiment analysis.However,these existing methods usually pursue high accuracy,ignoring the effect on model training efficiency.Considering th...Many methods based on deep learning have achieved amazing results in image sentiment analysis.However,these existing methods usually pursue high accuracy,ignoring the effect on model training efficiency.Considering that when faced with large-scale sentiment analysis tasks,the high accuracy rate often requires long experimental time.In view of the weakness,a method that can greatly improve experimental efficiency with only small fluctuations in model accuracy is proposed,and singular value decomposition(SVD)is used to find the sparse feature of the image,which are sparse vectors with strong discriminativeness and effectively reduce redundant information;The authors propose the Fast Dictionary Learning algorithm(FDL),which can combine neural network with sparse representation.This method is based on K-Singular Value Decomposition,and through iteration,it can effectively reduce the calculation time and greatly improve the training efficiency in the case of small fluctuation of accuracy.Moreover,the effectiveness of the proposed method is evaluated on the FER2013 dataset.By adding singular value decomposition,the accuracy of the test suite increased by 0.53%,and the total experiment time was shortened by 8.2%;Fast Dictionary Learning shortened the total experiment time by 36.3%.展开更多
Electronic medical records (EMR) facilitate the sharing of medical data, but existing sharing schemes suffer fromprivacy leakage and inefficiency. This article proposes a lightweight, searchable, and controllable EMR ...Electronic medical records (EMR) facilitate the sharing of medical data, but existing sharing schemes suffer fromprivacy leakage and inefficiency. This article proposes a lightweight, searchable, and controllable EMR sharingscheme, which employs a large attribute domain and a linear secret sharing structure (LSSS), the computationaloverhead of encryption and decryption reaches a lightweight constant level, and supports keyword search andpolicy hiding, which improves the high efficiency of medical data sharing. The dynamic accumulator technologyis utilized to enable data owners to flexibly authorize or revoke the access rights of data visitors to the datato achieve controllability of the data. Meanwhile, the data is re-encrypted by Intel Software Guard Extensions(SGX) technology to realize resistance to offline dictionary guessing attacks. In addition, blockchain technology isutilized to achieve credible accountability for abnormal behaviors in the sharing process. The experiments reflectthe obvious advantages of the scheme in terms of encryption and decryption computation overhead and storageoverhead, and theoretically prove the security and controllability in the sharing process, providing a feasible solutionfor the safe and efficient sharing of EMR.展开更多
Quantized training has been proven to be a prominent method to achieve deep neural network training under limited computational resources.It uses low bit-width arithmetics with a proper scaling factor to achieve negli...Quantized training has been proven to be a prominent method to achieve deep neural network training under limited computational resources.It uses low bit-width arithmetics with a proper scaling factor to achieve negligible accuracy loss.Cambricon-Q is the ASIC design proposed to efficiently support quantized training,and achieves significant performance improvement.However,there are still two caveats in the design.First,Cambricon-Q with different hardware specifications may lead to different numerical errors,resulting in non-reproducible behaviors which may become a major concern in critical applications.Second,Cambricon-Q cannot leverage data sparsity,where considerable cycles could still be squeezed out.To address the caveats,the acceleration core of Cambricon-Q is redesigned to support fine-grained irregular data processing.The new design not only enables acceleration on sparse data,but also enables performing local dynamic quantization by contiguous value ranges(which is hardware independent),instead of contiguous addresses(which is dependent on hardware factors).Experimental results show that the accuracy loss of the method still keeps negligible,and the accelerator achieves 1.61×performance improvement over Cambricon-Q,with about 10%energy increase.展开更多
The proportionate recursive least squares(PRLS)algorithm has shown faster convergence and better performance than both proportionate updating(PU)mechanism based least mean squares(LMS)algorithms and RLS algorithms wit...The proportionate recursive least squares(PRLS)algorithm has shown faster convergence and better performance than both proportionate updating(PU)mechanism based least mean squares(LMS)algorithms and RLS algorithms with a sparse regularization term.In this paper,we propose a variable forgetting factor(VFF)PRLS algorithm with a sparse penalty,e.g.,l_(1)-norm,for sparse identification.To reduce the computation complexity of the proposed algorithm,a fast implementation method based on dichotomous coordinate descent(DCD)algorithm is also derived.Simulation results indicate superior performance of the proposed algorithm.展开更多
In a cloud environment,outsourced graph data is widely used in companies,enterprises,medical institutions,and so on.Data owners and users can save costs and improve efficiency by storing large amounts of graph data on...In a cloud environment,outsourced graph data is widely used in companies,enterprises,medical institutions,and so on.Data owners and users can save costs and improve efficiency by storing large amounts of graph data on cloud servers.Servers on cloud platforms usually have some subjective or objective attacks,which make the outsourced graph data in an insecure state.The issue of privacy data protection has become an important obstacle to data sharing and usage.How to query outsourcing graph data safely and effectively has become the focus of research.Adjacency query is a basic and frequently used operation in graph,and it will effectively promote the query range and query ability if multi-keyword fuzzy search can be supported at the same time.This work proposes to protect the privacy information of outsourcing graph data by encryption,mainly studies the problem of multi-keyword fuzzy adjacency query,and puts forward a solution.In our scheme,we use the Bloom filter and encryption mechanism to build a secure index and query token,and adjacency queries are implemented through indexes and query tokens on the cloud server.Our proposed scheme is proved by formal analysis,and the performance and effectiveness of the scheme are illustrated by experimental analysis.The research results of this work will provide solid theoretical and technical support for the further popularization and application of encrypted graph data processing technology.展开更多
Cell migration plays a significant role in physiological and pathological processes.Understanding the characteristics of cell movement is crucial for comprehending biological processes such as cell functionality,cell ...Cell migration plays a significant role in physiological and pathological processes.Understanding the characteristics of cell movement is crucial for comprehending biological processes such as cell functionality,cell migration,and cell–cell interactions.One of the fundamental characteristics of cell movement is the specific distribution of cell speed,containing valuable information that still requires comprehensive understanding.This article investigates the distribution of mean velocities along cell trajectories,with a focus on optimizing the efficiency of cell food search in the context of the entire colony.We confirm that the specific velocity distribution in the experiments corresponds to an optimal search efficiency when spatial weighting is considered.The simulation results indicate that the distribution of average velocity does not align with the optimal search efficiency when employing average spatial weighting.However,when considering the distribution of central spatial weighting,the specific velocity distribution in the experiment is shown to correspond to the optimal search efficiency.Our simulations reveal that for any given distribution of average velocity,a specific central spatial weighting can be identified among the possible central spatial weighting that aligns with the optimal search strategy.Additionally,our work presents a method for determining the spatial weights embedded in the velocity distribution of cell movement.Our results have provided new avenues for further investigation of significant topics,such as relationship between cell behavior and environmental conditions throughout their evolutionary history,and how cells achieve collective cooperation through cell-cell communication.展开更多
Signal decomposition and multiscale signal analysis provide many useful tools for timefrequency analysis.We proposed a random feature method for analyzing time-series data by constructing a sparse approximation to the...Signal decomposition and multiscale signal analysis provide many useful tools for timefrequency analysis.We proposed a random feature method for analyzing time-series data by constructing a sparse approximation to the spectrogram.The randomization is both in the time window locations and the frequency sampling,which lowers the overall sampling and computational cost.The sparsification of the spectrogram leads to a sharp separation between time-frequency clusters which makes it easier to identify intrinsic modes,and thus leads to a new data-driven mode decomposition.The applications include signal representation,outlier removal,and mode decomposition.On benchmark tests,we show that our approach outperforms other state-of-the-art decomposition methods.展开更多
基金the University of Transport Technology under grant number DTTD2022-12.
文摘Determination of Shear Bond strength(SBS)at interlayer of double-layer asphalt concrete is crucial in flexible pavement structures.The study used three Machine Learning(ML)models,including K-Nearest Neighbors(KNN),Extra Trees(ET),and Light Gradient Boosting Machine(LGBM),to predict SBS based on easily determinable input parameters.Also,the Grid Search technique was employed for hyper-parameter tuning of the ML models,and cross-validation and learning curve analysis were used for training the models.The models were built on a database of 240 experimental results and three input variables:temperature,normal pressure,and tack coat rate.Model validation was performed using three statistical criteria:the coefficient of determination(R2),the Root Mean Square Error(RMSE),and the mean absolute error(MAE).Additionally,SHAP analysis was also used to validate the importance of the input variables in the prediction of the SBS.Results show that these models accurately predict SBS,with LGBM providing outstanding performance.SHAP(Shapley Additive explanation)analysis for LGBM indicates that temperature is the most influential factor on SBS.Consequently,the proposed ML models can quickly and accurately predict SBS between two layers of asphalt concrete,serving practical applications in flexible pavement structure design.
基金supported by the Scientific Research Project of Xiang Jiang Lab(22XJ02003)the University Fundamental Research Fund(23-ZZCX-JDZ-28)+5 种基金the National Science Fund for Outstanding Young Scholars(62122093)the National Natural Science Foundation of China(72071205)the Hunan Graduate Research Innovation Project(ZC23112101-10)the Hunan Natural Science Foundation Regional Joint Project(2023JJ50490)the Science and Technology Project for Young and Middle-aged Talents of Hunan(2023TJ-Z03)the Science and Technology Innovation Program of Humnan Province(2023RC1002)。
文摘Traditional large-scale multi-objective optimization algorithms(LSMOEAs)encounter difficulties when dealing with sparse large-scale multi-objective optimization problems(SLM-OPs)where most decision variables are zero.As a result,many algorithms use a two-layer encoding approach to optimize binary variable Mask and real variable Dec separately.Nevertheless,existing optimizers often focus on locating non-zero variable posi-tions to optimize the binary variables Mask.However,approxi-mating the sparse distribution of real Pareto optimal solutions does not necessarily mean that the objective function is optimized.In data mining,it is common to mine frequent itemsets appear-ing together in a dataset to reveal the correlation between data.Inspired by this,we propose a novel two-layer encoding learning swarm optimizer based on frequent itemsets(TELSO)to address these SLMOPs.TELSO mined the frequent terms of multiple particles with better target values to find mask combinations that can obtain better objective values for fast convergence.Experi-mental results on five real-world problems and eight benchmark sets demonstrate that TELSO outperforms existing state-of-the-art sparse large-scale multi-objective evolutionary algorithms(SLMOEAs)in terms of performance and convergence speed.
文摘Passive detection of low-slow-small(LSS)targets is easily interfered by direct signal and multipath clutter,and the traditional clutter suppression method has the contradiction between step size and convergence rate.In this paper,a frequency domain clutter suppression algorithm based on sparse adaptive filtering is proposed.The pulse compression operation between the error signal and the input reference signal is added to the cost function as a sparsity constraint,and the criterion for filter weight updating is improved to obtain a purer echo signal.At the same time,the step size and penalty factor are brought into the adaptive iteration process,and the input data is used to drive the adaptive changes of parameters such as step size.The proposed algorithm has a small amount of calculation,which improves the robustness to parameters such as step size,reduces the weight error of the filter and has a good clutter suppression performance.
基金supported in part by the National Natural Science Foundation of China (NSFC) under Grant No.61976242in part by the Natural Science Fund of Hebei Province for Distinguished Young Scholars under Grant No.F2021202010+2 种基金in part by the Fundamental Scientific Research Funds for Interdisciplinary Team of Hebei University of Technology under Grant No.JBKYTD2002funded by Science and Technology Project of Hebei Education Department under Grant No.JZX2023007supported by 2022 Interdisciplinary Postgraduate Training Program of Hebei University of Technology under Grant No.HEBUT-YXKJC-2022122.
文摘Most of the neural network architectures are based on human experience,which requires a long and tedious trial-and-error process.Neural architecture search(NAS)attempts to detect effective architectures without human intervention.Evolutionary algorithms(EAs)for NAS can find better solutions than human-designed architectures by exploring a large search space for possible architectures.Using multiobjective EAs for NAS,optimal neural architectures that meet various performance criteria can be explored and discovered efficiently.Furthermore,hardware-accelerated NAS methods can improve the efficiency of the NAS.While existing reviews have mainly focused on different strategies to complete NAS,a few studies have explored the use of EAs for NAS.In this paper,we summarize and explore the use of EAs for NAS,as well as large-scale multiobjective optimization strategies and hardware-accelerated NAS methods.NAS performs well in healthcare applications,such as medical image analysis,classification of disease diagnosis,and health monitoring.EAs for NAS can automate the search process and optimize multiple objectives simultaneously in a given healthcare task.Deep neural network has been successfully used in healthcare,but it lacks interpretability.Medical data is highly sensitive,and privacy leaks are frequently reported in the healthcare industry.To solve these problems,in healthcare,we propose an interpretable neuroevolution framework based on federated learning to address search efficiency and privacy protection.Moreover,we also point out future research directions for evolutionary NAS.Overall,for researchers who want to use EAs to optimize NNs in healthcare,we analyze the advantages and disadvantages of doing so to provide detailed guidance,and propose an interpretable privacy-preserving framework for healthcare applications.
基金the support of the National Natural Science Foundation of China(Grant No.62076204)the Seed Foundation of Innovation and Creation for Graduate Students in Northwestern Polytechnical University(Grant No.CX2020019)in part by the China Postdoctoral Science Foundation(Grants No.2021M700337)。
文摘Improvement of integrated battlefield situational awareness in complex environments involving dynamic factors such as restricted communications and electromagnetic interference(EMI)has become a contentious research problem.In certain mission environments,due to the impact of many interference sources on real-time communication or mission requirements such as the need to implement communication regulations,the mission stages are represented as a dynamic combination of several communication-available and communication-unavailable stages.Furthermore,the data interaction between unmanned aerial vehicles(UAVs)can only be performed in specific communication-available stages.Traditional cooperative search algorithms cannot handle such situations well.To solve this problem,this study constructed a distributed model predictive control(DMPC)architecture for a collaborative control of UAVs and used the Voronoi diagram generation method to re-plan the search areas of all UAVs in real time to avoid repetition of search areas and UAV collisions while improving the search efficiency and safety factor.An attention mechanism ant-colony optimization(AACO)algorithm is proposed for UAV search-control decision planning.The search strategy is adaptively updated by introducing an attention mechanism for regular instruction information,a priori information,and emergent information of the mission to satisfy different search expectations to the maximum extent.Simulation results show that the proposed algorithm achieves better search performance than traditional algorithms in restricted communication constraint scenarios.
基金support by the Open Project of Xiangjiang Laboratory(22XJ02003)the University Fundamental Research Fund(23-ZZCX-JDZ-28,ZK21-07)+5 种基金the National Science Fund for Outstanding Young Scholars(62122093)the National Natural Science Foundation of China(72071205)the Hunan Graduate Research Innovation Project(CX20230074)the Hunan Natural Science Foundation Regional Joint Project(2023JJ50490)the Science and Technology Project for Young and Middle-aged Talents of Hunan(2023TJZ03)the Science and Technology Innovation Program of Humnan Province(2023RC1002).
文摘Sparse large-scale multi-objective optimization problems(SLMOPs)are common in science and engineering.However,the large-scale problem represents the high dimensionality of the decision space,requiring algorithms to traverse vast expanse with limited computational resources.Furthermore,in the context of sparse,most variables in Pareto optimal solutions are zero,making it difficult for algorithms to identify non-zero variables efficiently.This paper is dedicated to addressing the challenges posed by SLMOPs.To start,we introduce innovative objective functions customized to mine maximum and minimum candidate sets.This substantial enhancement dramatically improves the efficacy of frequent pattern mining.In this way,selecting candidate sets is no longer based on the quantity of nonzero variables they contain but on a higher proportion of nonzero variables within specific dimensions.Additionally,we unveil a novel approach to association rule mining,which delves into the intricate relationships between non-zero variables.This novel methodology aids in identifying sparse distributions that can potentially expedite reductions in the objective function value.We extensively tested our algorithm across eight benchmark problems and four real-world SLMOPs.The results demonstrate that our approach achieves competitive solutions across various challenges.
文摘Aiming at the practical application of Unmanned Underwater Vehicle(UUV)in underwater combat,this paper proposes a battlefield ambush scene with UUV considering ocean current.Firstly,by establishing these mathematical models of ocean current environment,target movement,and sonar detection,the probability calculation methods of single UUV searching target and multiple UUV cooperatively searching target are given respectively.Then,based on the Hybrid Quantum-behaved Particle Swarm Optimization(HQPSO)algorithm,the path with the highest target search probability is found.Finally,through simulation calculations,the influence of different UUV parameters and target parameters on the target search probability is analyzed,and the minimum number of UUVs that need to be deployed to complete the ambush task is demonstrated,and the optimal search path scheme is obtained.The method proposed in this paper provides a theoretical basis for the practical application of UUV in the future combat.
基金This work was supported in part by the National Natural Science Foundation of China under Grant 61305001the Natural Science Foundation of Heilongjiang Province of China under Grant F201222.
文摘In differentiable search architecture search methods,a more efficient search space design can significantly improve the performance of the searched architecture,thus requiring people to carefully define the search space with different complexity according to various operations.Meanwhile rationalizing the search strategies to explore the well-defined search space will further improve the speed and efficiency of architecture search.With this in mind,we propose a faster and more efficient differentiable architecture search method,AllegroNAS.Firstly,we introduce a more efficient search space enriched by the introduction of two redefined convolution modules.Secondly,we utilize a more efficient architectural parameter regularization method,mitigating the overfitting problem during the search process and reducing the error brought about by gradient approximation.Meanwhile,we introduce a natural exponential cosine annealing method to make the learning rate of the neural network training process more suitable for the search procedure.Moreover,group convolution and data augmentation are employed to reduce the computational cost.Finally,through extensive experiments on several public datasets,we demonstrate that our method can more swiftly search for better-performing neural network architectures in a more efficient search space,thus validating the effectiveness of our approach.
文摘Contract Bridge,a four-player imperfect information game,comprises two phases:bidding and playing.While computer programs excel at playing,bidding presents a challenging aspect due to the need for information exchange with partners and interference with communication of opponents.In this work,we introduce a Bridge bidding agent that combines supervised learning,deep reinforcement learning via self-play,and a test-time search approach.Our experiments demonstrate that our agent outperforms WBridge5,a highly regarded computer Bridge software that has won multiple world championships,by a performance of 0.98 IMPs(international match points)per deal over 10000 deals,with a much cost-effective approach.The performance significantly surpasses previous state-of-the-art(0.85 IMPs per deal).Note 0.1 IMPs per deal is a significant improvement in Bridge bidding.
基金supported by Natural Science Foundation of China(62071262)the K.C.Wong Magna Fund at Ningbo University.
文摘In this paper,we reconstruct strongly-decaying block sparse signals by the block generalized orthogonal matching pursuit(BgOMP)algorithm in the l2-bounded noise case.Under some restraints on the minimum magnitude of the nonzero elements of the strongly-decaying block sparse signal,if the sensing matrix satisfies the the block restricted isometry property(block-RIP),then arbitrary strongly-decaying block sparse signals can be accurately and steadily reconstructed by the BgOMP algorithm in iterations.Furthermore,we conjecture that this condition is sharp.
基金supported in part by NUS startup grantthe National Natural Science Foundation of China (52076037)。
文摘Although many multi-view clustering(MVC) algorithms with acceptable performances have been presented, to the best of our knowledge, nearly all of them need to be fed with the correct number of clusters. In addition, these existing algorithms create only the hard and fuzzy partitions for multi-view objects,which are often located in highly-overlapping areas of multi-view feature space. The adoption of hard and fuzzy partition ignores the ambiguity and uncertainty in the assignment of objects, likely leading to performance degradation. To address these issues, we propose a novel sparse reconstructive multi-view evidential clustering algorithm(SRMVEC). Based on a sparse reconstructive procedure, SRMVEC learns a shared affinity matrix across views, and maps multi-view objects to a 2-dimensional humanreadable chart by calculating 2 newly defined mathematical metrics for each object. From this chart, users can detect the number of clusters and select several objects existing in the dataset as cluster centers. Then, SRMVEC derives a credal partition under the framework of evidence theory, improving the fault tolerance of clustering. Ablation studies show the benefits of adopting the sparse reconstructive procedure and evidence theory. Besides,SRMVEC delivers effectiveness on benchmark datasets by outperforming some state-of-the-art methods.
基金supported by the National Natural Science Foundation of China(No.61801440)the High‐quality and Cutting‐edge Disciplines Construction Project for Universities in Beijing(Internet Information,Communication University of China),State Key Laboratory of Media Convergence and Communication(Communication University of China)the Fundamental Research Funds for the Central Universities(CUC2019B069).
文摘Many methods based on deep learning have achieved amazing results in image sentiment analysis.However,these existing methods usually pursue high accuracy,ignoring the effect on model training efficiency.Considering that when faced with large-scale sentiment analysis tasks,the high accuracy rate often requires long experimental time.In view of the weakness,a method that can greatly improve experimental efficiency with only small fluctuations in model accuracy is proposed,and singular value decomposition(SVD)is used to find the sparse feature of the image,which are sparse vectors with strong discriminativeness and effectively reduce redundant information;The authors propose the Fast Dictionary Learning algorithm(FDL),which can combine neural network with sparse representation.This method is based on K-Singular Value Decomposition,and through iteration,it can effectively reduce the calculation time and greatly improve the training efficiency in the case of small fluctuation of accuracy.Moreover,the effectiveness of the proposed method is evaluated on the FER2013 dataset.By adding singular value decomposition,the accuracy of the test suite increased by 0.53%,and the total experiment time was shortened by 8.2%;Fast Dictionary Learning shortened the total experiment time by 36.3%.
基金the Natural Science Foundation of Hebei Province under Grant Number F2021201052.
文摘Electronic medical records (EMR) facilitate the sharing of medical data, but existing sharing schemes suffer fromprivacy leakage and inefficiency. This article proposes a lightweight, searchable, and controllable EMR sharingscheme, which employs a large attribute domain and a linear secret sharing structure (LSSS), the computationaloverhead of encryption and decryption reaches a lightweight constant level, and supports keyword search andpolicy hiding, which improves the high efficiency of medical data sharing. The dynamic accumulator technologyis utilized to enable data owners to flexibly authorize or revoke the access rights of data visitors to the datato achieve controllability of the data. Meanwhile, the data is re-encrypted by Intel Software Guard Extensions(SGX) technology to realize resistance to offline dictionary guessing attacks. In addition, blockchain technology isutilized to achieve credible accountability for abnormal behaviors in the sharing process. The experiments reflectthe obvious advantages of the scheme in terms of encryption and decryption computation overhead and storageoverhead, and theoretically prove the security and controllability in the sharing process, providing a feasible solutionfor the safe and efficient sharing of EMR.
基金the National Key Research and Devecopment Program of China(No.2022YFB4501601)the National Natural Science Foundation of China(No.62102398,U20A20227,62222214,62002338,U22A2028,U19B2019)+1 种基金the Chinese Academy of Sciences Project for Young Scientists in Basic Research(YSBR-029)Youth Innovation Promotion Association Chinese Academy of Sciences。
文摘Quantized training has been proven to be a prominent method to achieve deep neural network training under limited computational resources.It uses low bit-width arithmetics with a proper scaling factor to achieve negligible accuracy loss.Cambricon-Q is the ASIC design proposed to efficiently support quantized training,and achieves significant performance improvement.However,there are still two caveats in the design.First,Cambricon-Q with different hardware specifications may lead to different numerical errors,resulting in non-reproducible behaviors which may become a major concern in critical applications.Second,Cambricon-Q cannot leverage data sparsity,where considerable cycles could still be squeezed out.To address the caveats,the acceleration core of Cambricon-Q is redesigned to support fine-grained irregular data processing.The new design not only enables acceleration on sparse data,but also enables performing local dynamic quantization by contiguous value ranges(which is hardware independent),instead of contiguous addresses(which is dependent on hardware factors).Experimental results show that the accuracy loss of the method still keeps negligible,and the accelerator achieves 1.61×performance improvement over Cambricon-Q,with about 10%energy increase.
基金supported by National Key Research and Development Program of China(2020YFB0505803)National Key Research and Development Program of China(2016YFB0501700)。
文摘The proportionate recursive least squares(PRLS)algorithm has shown faster convergence and better performance than both proportionate updating(PU)mechanism based least mean squares(LMS)algorithms and RLS algorithms with a sparse regularization term.In this paper,we propose a variable forgetting factor(VFF)PRLS algorithm with a sparse penalty,e.g.,l_(1)-norm,for sparse identification.To reduce the computation complexity of the proposed algorithm,a fast implementation method based on dichotomous coordinate descent(DCD)algorithm is also derived.Simulation results indicate superior performance of the proposed algorithm.
基金This research was supported in part by the Nature Science Foundation of China(Nos.62262033,61962029,61762055,62062045 and 62362042)the Jiangxi Provincial Natural Science Foundation of China(Nos.20224BAB202012,20202ACBL202005 and 20202BAB212006)+3 种基金the Science and Technology Research Project of Jiangxi Education Department(Nos.GJJ211815,GJJ2201914 and GJJ201832)the Hubei Natural Science Foundation Innovation and Development Joint Fund Project(No.2022CFD101)Xiangyang High-Tech Key Science and Technology Plan Project(No.2022ABH006848)Hubei Superior and Distinctive Discipline Group of“New Energy Vehicle and Smart Transportation”,the Project of Zhejiang Institute of Mechanical&Electrical Engineering,and the Jiangxi Provincial Social Science Foundation of China(No.23GL52D).
文摘In a cloud environment,outsourced graph data is widely used in companies,enterprises,medical institutions,and so on.Data owners and users can save costs and improve efficiency by storing large amounts of graph data on cloud servers.Servers on cloud platforms usually have some subjective or objective attacks,which make the outsourced graph data in an insecure state.The issue of privacy data protection has become an important obstacle to data sharing and usage.How to query outsourcing graph data safely and effectively has become the focus of research.Adjacency query is a basic and frequently used operation in graph,and it will effectively promote the query range and query ability if multi-keyword fuzzy search can be supported at the same time.This work proposes to protect the privacy information of outsourcing graph data by encryption,mainly studies the problem of multi-keyword fuzzy adjacency query,and puts forward a solution.In our scheme,we use the Bloom filter and encryption mechanism to build a secure index and query token,and adjacency queries are implemented through indexes and query tokens on the cloud server.Our proposed scheme is proved by formal analysis,and the performance and effectiveness of the scheme are illustrated by experimental analysis.The research results of this work will provide solid theoretical and technical support for the further popularization and application of encrypted graph data processing technology.
基金Project supported by the National Natural Science Foundation of China(Grant No.31971183).
文摘Cell migration plays a significant role in physiological and pathological processes.Understanding the characteristics of cell movement is crucial for comprehending biological processes such as cell functionality,cell migration,and cell–cell interactions.One of the fundamental characteristics of cell movement is the specific distribution of cell speed,containing valuable information that still requires comprehensive understanding.This article investigates the distribution of mean velocities along cell trajectories,with a focus on optimizing the efficiency of cell food search in the context of the entire colony.We confirm that the specific velocity distribution in the experiments corresponds to an optimal search efficiency when spatial weighting is considered.The simulation results indicate that the distribution of average velocity does not align with the optimal search efficiency when employing average spatial weighting.However,when considering the distribution of central spatial weighting,the specific velocity distribution in the experiment is shown to correspond to the optimal search efficiency.Our simulations reveal that for any given distribution of average velocity,a specific central spatial weighting can be identified among the possible central spatial weighting that aligns with the optimal search strategy.Additionally,our work presents a method for determining the spatial weights embedded in the velocity distribution of cell movement.Our results have provided new avenues for further investigation of significant topics,such as relationship between cell behavior and environmental conditions throughout their evolutionary history,and how cells achieve collective cooperation through cell-cell communication.
基金supported in part by the NSERC RGPIN 50503-10842supported in part by the AFOSR MURI FA9550-21-1-0084the NSF DMS-1752116.
文摘Signal decomposition and multiscale signal analysis provide many useful tools for timefrequency analysis.We proposed a random feature method for analyzing time-series data by constructing a sparse approximation to the spectrogram.The randomization is both in the time window locations and the frequency sampling,which lowers the overall sampling and computational cost.The sparsification of the spectrogram leads to a sharp separation between time-frequency clusters which makes it easier to identify intrinsic modes,and thus leads to a new data-driven mode decomposition.The applications include signal representation,outlier removal,and mode decomposition.On benchmark tests,we show that our approach outperforms other state-of-the-art decomposition methods.