The pursuit-evasion game models the strategic interaction among players, attracting attention in many realistic scenarios, such as missile guidance, unmanned aerial vehicles, and target defense. Existing studies mainl...The pursuit-evasion game models the strategic interaction among players, attracting attention in many realistic scenarios, such as missile guidance, unmanned aerial vehicles, and target defense. Existing studies mainly concentrate on the cooperative pursuit of multiple players in two-dimensional pursuit-evasion games. However, these approaches can hardly be applied to practical situations where players usually move in three-dimensional space with a three-degree-of-freedom control. In this paper,we make the first attempt to investigate the equilibrium strategy of the realistic pursuit-evasion game, in which the pursuer follows a three-degree-of-freedom control, and the evader moves freely. First, we describe the pursuer's three-degree-of-freedom control and the evader's relative coordinate. We then rigorously derive the equilibrium strategy by solving the retrogressive path equation according to the Hamilton-Jacobi-Bellman-Isaacs(HJBI) method, which divides the pursuit-evasion process into the navigation and acceleration phases. Besides, we analyze the maximum allowable speed for the pursuer to capture the evader successfully and provide the strategy with which the evader can escape when the pursuer's speed exceeds the threshold. We further conduct comparison tests with various unilateral deviations to verify that the proposed strategy forms a Nash equilibrium.展开更多
Current successes in artificial intelligence domain have revitalized interest in spacecraft pursuit-evasion game,which is an interception problem with a non-cooperative maneuvering target.The paper presents an automat...Current successes in artificial intelligence domain have revitalized interest in spacecraft pursuit-evasion game,which is an interception problem with a non-cooperative maneuvering target.The paper presents an automated machine learning(AutoML)based method to generate optimal trajectories in long-distance scenarios.Compared with conventional deep neural network(DNN)methods,the proposed method dramatically reduces the reliance on manual intervention and machine learning expertise.Firstly,based on differential game theory and costate normalization technique,the trajectory optimization problem is formulated under the assumption of continuous thrust.Secondly,the AutoML technique based on sequential model-based optimization(SMBO)framework is introduced to automate DNN design in deep learning process.If recommended DNN architecture exists,the tree-structured Parzen estimator(TPE)is used,otherwise the efficient neural architecture search(NAS)with network morphism is used.Thus,a novel trajectory optimization method with high computational efficiency is achieved.Finally,numerical results demonstrate the feasibility and efficiency of the proposed method.展开更多
With the development of space rendezvous and proximity operations(RPO)in recent years,the scenarios with noncooperative spacecraft are attracting the attention of more and more researchers.A method based on the costat...With the development of space rendezvous and proximity operations(RPO)in recent years,the scenarios with noncooperative spacecraft are attracting the attention of more and more researchers.A method based on the costate normalization technique and deep neural networks is presented to generate the optimal guidance law for free-time orbital pursuit-evasion game.Firstly,the 24-dimensional problem given by differential game theory is transformed into a three-parameter optimization problem through the dimension-reduction method which guarantees the uniqueness of solution for the specific scenario.Secondly,a close-loop interactive mechanism involving feedback is introduced to deep neural networks for generating precise initial solution.Thus the optimal guidance law is obtained efficiently and stably with the application of optimization algorithm initialed by the deep neural networks.Finally,the results of the comparison with another two methods and Monte Carlo simulation demonstrate the efficiency and robustness of the proposed optimal guidance method.展开更多
In this paper,the pursuit-evasion game with state and control constraints is solved to achieve the Nash equilibrium of both the pursuer and the evader with an iterative self-play technique.Under the condition where th...In this paper,the pursuit-evasion game with state and control constraints is solved to achieve the Nash equilibrium of both the pursuer and the evader with an iterative self-play technique.Under the condition where the Hamiltonian formed by means of Pontryagin’s maximum principle has the unique solution,it can be proven that the iterative control law converges to the Nash equilibrium solution.However,the strong nonlinearity of the ordinary differential equations formulated by Pontryagin’s maximum principle makes the control policy difficult to figured out.Moreover the system dynamics employed in this manuscript contains a high dimensional state vector with constraints.In practical applications,such as the control of aircraft,the provided overload is limited.Therefore,in this paper,we consider the optimal strategy of pursuit-evasion games with constant constraint on the control,while some state vectors are restricted by the function of the input.To address the challenges,the optimal control problems are transformed into nonlinear programming problems through the direct collocation method.Finally,two numerical cases of the aircraft pursuit-evasion scenario are given to demonstrate the effectiveness of the presented method to obtain the optimal control of both the pursuer and the evader.展开更多
This work is inspired by a stealth pursuit behavior called motion camouflage whereby a pursuer approaches an evader while the pursuer camouflages itself against a predetermined background.We formulate the spacecraft p...This work is inspired by a stealth pursuit behavior called motion camouflage whereby a pursuer approaches an evader while the pursuer camouflages itself against a predetermined background.We formulate the spacecraft pursuit-evasion problem as a stealth pursuit strategy of motion camouflage,in which the pursuer tries to minimize a motion camouflage index defined in this paper.The Euler-Hill reference frame whose origin is set on the circular reference orbit is used to describe the dynamics.Based on the rule of motion camouflage,a guidance strategy in open-loop form to achieve motion camouflage index is derived in which the pursuer lies on the camouflage constraint line connecting the central spacecraft and evader.In order to dispose of the dependence on the evader acceleration in the open-loop guidance strategy,we further consider the motion camouflage pursuit problem within an infinite-horizon nonlinear quadratic differential game.The saddle point solution to the game is derived by using the state-dependent Riccati equation method,and the resulting closed-loop guidance strategy is effective in achieving motion camouflage.Simulations are performed to demonstrate the capabilities of the proposed guidance strategies for the pursuit–evasion game scenario.展开更多
TheUAV pursuit-evasion problem focuses on the efficient tracking and capture of evading targets using unmanned aerial vehicles(UAVs),which is pivotal in public safety applications,particularly in scenarios involving i...TheUAV pursuit-evasion problem focuses on the efficient tracking and capture of evading targets using unmanned aerial vehicles(UAVs),which is pivotal in public safety applications,particularly in scenarios involving intrusion monitoring and interception.To address the challenges of data acquisition,real-world deployment,and the limited intelligence of existing algorithms in UAV pursuit-evasion tasks,we propose an innovative swarm intelligencebased UAV pursuit-evasion control framework,namely“Boids Model-based DRL Approach for Pursuit and Escape”(Boids-PE),which synergizes the strengths of swarm intelligence from bio-inspired algorithms and deep reinforcement learning(DRL).The Boids model,which simulates collective behavior through three fundamental rules,separation,alignment,and cohesion,is adopted in our work.By integrating Boids model with the Apollonian Circles algorithm,significant improvements are achieved in capturing UAVs against simple evasion strategies.To further enhance decision-making precision,we incorporate a DRL algorithm to facilitate more accurate strategic planning.We also leverage self-play training to continuously optimize the performance of pursuit UAVs.During experimental evaluation,we meticulously designed both one-on-one and multi-to-one pursuit-evasion scenarios,customizing the state space,action space,and reward function models for each scenario.Extensive simulations,supported by the PyBullet physics engine,validate the effectiveness of our proposed method.The overall results demonstrate that Boids-PE significantly enhance the efficiency and reliability of UAV pursuit-evasion tasks,providing a practical and robust solution for the real-world application of UAV pursuit-evasion missions.展开更多
The orbital pursuit-evasion game is typically formulated as a complete-information game,which assumes the payoff functions of the two players are common knowledge.However,realistic pursuit-evasion games typically have...The orbital pursuit-evasion game is typically formulated as a complete-information game,which assumes the payoff functions of the two players are common knowledge.However,realistic pursuit-evasion games typically have incomplete information,in which the lack of payoff information limits the player’s ability to play optimally.To address this problem,this paper proposes a currently optimal escape strategy based on estimation for the evader.In this strategy,the currently optimal evasive controls are first derived based on the evader’s guess of the pursuer’s payoff weightings.Then an online parameter estimation method based on a modified strong tracking unscented Kalman filter is employed to modify the guess and update the strategy during the game.As the estimation becomes accurate,the currently optimal strategy gets closer to the actually optimal strategy.Simulation results show the proposed strategy can achieve optimal evasive controls progressively and the evader’s payoff of the strategy is lower than that of the zero-sum escape strategy.Meanwhile,the proposed strategy is also effective in the case where the pursuer changes its payoff function halfway during the game.展开更多
In practical combat scenario,the cooperative intercept strategies are often carefully designed,and it is challenging for the hypersonic vehicles to achieve successful evasion.Based on the analysis,it can be found that...In practical combat scenario,the cooperative intercept strategies are often carefully designed,and it is challenging for the hypersonic vehicles to achieve successful evasion.Based on the analysis,it can be found that if several Successive Pursuers come from the Same Direction(SPSD)and flight with a proper spacing,the evasion difficulty may increase greatly.To address this problem,we focus on the evasion guidance strategy design for the Air-breathing Hypersonic Vehicles(AHVs)under the SPSD combat scenario.In order to avoid the induced influence on the scramjet,altitude and speed of the vehicle,the lateral maneuver and evasion are employed.To guarantee the remnant maneuver ability,the concept of specified miss distance is introduced and utilized to generate the guidance command for the AHV.In the framework of constrained optimal control,the analytical expression of the evasion command is derived,and the constraints of the overload can be ensured to be never violated.In fact,by analyzing the spacing of the pursers,it can be classified whether the cooperative pursuit is formed.For the coordination-unformed multiple pursers,the evasion can be achieved lightly by the proposed strategy.If the coordination is formed,the proposed method will generate a large reverse direction maneuver,and the successful evasion can be maintained as a result.The performance of the proposed algorithms is tested in numerical simulations.展开更多
To improve the anti-jamming and interference mitigation ability of the UAV-aided communication systems, this paper investigates the channel selection optimization problem in face of both internal mutual interference a...To improve the anti-jamming and interference mitigation ability of the UAV-aided communication systems, this paper investigates the channel selection optimization problem in face of both internal mutual interference and external malicious jamming. A cooperative anti-jamming and interference mitigation method based on local altruistic is proposed to optimize UAVs’ channel selection. Specifically, a Stackelberg game is modeled to formulate the confrontation relationship between UAVs and the jammer. A local altruistic game is modeled with each UAV considering the utilities of both itself and other UAVs. A distributed cooperative anti-jamming and interference mitigation algorithm is proposed to obtain the Stackelberg equilibrium. Finally, the convergence of the proposed algorithm and the impact of the transmission power on the system loss value are analyzed, and the anti-jamming performance of the proposed algorithm can be improved by around 64% compared with the existing algorithms.展开更多
Benefiting from the development of Federated Learning(FL)and distributed communication systems,large-scale intelligent applications become possible.Distributed devices not only provide adequate training data,but also ...Benefiting from the development of Federated Learning(FL)and distributed communication systems,large-scale intelligent applications become possible.Distributed devices not only provide adequate training data,but also cause privacy leakage and energy consumption.How to optimize the energy consumption in distributed communication systems,while ensuring the privacy of users and model accuracy,has become an urgent challenge.In this paper,we define the FL as a 3-layer architecture including users,agents and server.In order to find a balance among model training accuracy,privacy-preserving effect,and energy consumption,we design the training process of FL as game models.We use an extensive game tree to analyze the key elements that influence the players’decisions in the single game,and then find the incentive mechanism that meet the social norms through the repeated game.The experimental results show that the Nash equilibrium we obtained satisfies the laws of reality,and the proposed incentive mechanism can also promote users to submit high-quality data in FL.Following the multiple rounds of play,the incentive mechanism can help all players find the optimal strategies for energy,privacy,and accuracy of FL in distributed communication systems.展开更多
Malicious attacks against data are unavoidable in the interconnected,open and shared Energy Internet(EI),Intrusion tolerant techniques are critical to the data security of EI.Existing intrusion tolerant techniques suf...Malicious attacks against data are unavoidable in the interconnected,open and shared Energy Internet(EI),Intrusion tolerant techniques are critical to the data security of EI.Existing intrusion tolerant techniques suffered from problems such as low adaptability,policy lag,and difficulty in determining the degree of tolerance.To address these issues,we propose a novel adaptive intrusion tolerance model based on game theory that enjoys two-fold ideas:(1)it constructs an improved replica of the intrusion tolerance model of the dynamic equation evolution game to induce incentive weights;and (2)it combines a tournament competition model with incentive weights to obtain optimal strategies for each stage of the game process.Extensive experiments are conducted in the IEEE 39-bus system,whose results demonstrate the feasibility of the incentive weights,confirm the proposed strategy strengthens the system’s ability to tolerate aggression,and improves the dynamic adaptability and response efficiency of the aggression-tolerant system in the case of limited resources.展开更多
1.Introduction In August 2024,over 4400 Paralympic athletes will gather in Paris for the Paralympic Summer Games—the pinnacle of every Paralympian’s(Para athletes competing at the Paralympic Games)career to showcase...1.Introduction In August 2024,over 4400 Paralympic athletes will gather in Paris for the Paralympic Summer Games—the pinnacle of every Paralympian’s(Para athletes competing at the Paralympic Games)career to showcase their ability and skills.Their training,preparation,and effort in the years leading up to the Games are unparalleled.To achieve success,Paralympians specifically rely on a medical support team to achieve their goals.So,what is required of the medical support team to prepare Paralympians to get ready,set,and go to Paris 2024?展开更多
Objective: To study the problematic use of video games among secondary school students in the city of Parakou in 2023. Methods: Descriptive cross-sectional study conducted in the commune of Parakou from December 2022 ...Objective: To study the problematic use of video games among secondary school students in the city of Parakou in 2023. Methods: Descriptive cross-sectional study conducted in the commune of Parakou from December 2022 to July 2023. The study population consisted of students regularly enrolled in public and private secondary schools in the city of Parakou for the 2022-2023 academic year. A two-stage non-proportional stratified sampling technique combined with simple random sampling was adopted. The Problem Video Game Playing (PVP) scale was used to assess problem gambling in the study population, while anxiety and depression were assessed using the Hospital Anxiety and Depression Scale (HADS). Results: A total of 1030 students were included. The mean age of the pupils surveyed was 15.06 ± 2.68 years, with extremes of 10 and 28 years. The [13 - 18] age group was the most represented, with a proportion of 59.6% (614) in the general population. Females predominated, at 52.8% (544), with a sex ratio of 0.89. The prevalence of problematic video game use was 24.9%, measured using the Video Game Playing scale. Associated factors were male gender (p = 0.005), pocket money under 10,000 cfa (p = 0.001) and between 20,000 - 90,000 cfa (p = 0.030), addictive family behavior (p < 0.001), monogamous family (p = 0.023), good relationship with father (p = 0.020), organization of video game competitions (p = 0.001) and definite anxiety (p Conclusion: Substance-free addiction is struggling to attract the attention it deserves, as it did in its infancy everywhere else. This study complements existing data and serves as a reminder of the need to focus on this group of addictions, whose problematic use of video games remains the most frequent due to its accessibility and social tolerance. Preventive action combined with curative measures remains the most effective means of combating the problem at national level.展开更多
Given a graph g=( V,A ) , we define a space of subgraphs M with the binary operation of union and the unique decomposition property into blocks. This space allows us to discuss a notion of minimal subgraphs (minimal c...Given a graph g=( V,A ) , we define a space of subgraphs M with the binary operation of union and the unique decomposition property into blocks. This space allows us to discuss a notion of minimal subgraphs (minimal coalitions) that are of interest for the game. Additionally, a partition of the game is defined in terms of the gain of each block, and subsequently, a solution to the game is defined based on distributing to each player (node and edge) present in each block a payment proportional to their contribution to the coalition.展开更多
In public goods games, punishments and rewards have been shown to be effective mechanisms for maintaining individualcooperation. However, punishments and rewards are costly to incentivize cooperation. Therefore, the g...In public goods games, punishments and rewards have been shown to be effective mechanisms for maintaining individualcooperation. However, punishments and rewards are costly to incentivize cooperation. Therefore, the generation ofcostly penalties and rewards has been a complex problem in promoting the development of cooperation. In real society,specialized institutions exist to punish evil people or reward good people by collecting taxes. We propose a strong altruisticpunishment or reward strategy in the public goods game through this phenomenon. Through theoretical analysis and numericalcalculation, we can get that tax-based strong altruistic punishment (reward) has more evolutionary advantages thantraditional strong altruistic punishment (reward) in maintaining cooperation and tax-based strong altruistic reward leads toa higher level of cooperation than tax-based strong altruistic punishment.展开更多
As the current global environment is deteriorating,distributed renewable energy is gradually becoming an important member of the energy internet.Blockchain,as a decentralized distributed ledger with decentralization,t...As the current global environment is deteriorating,distributed renewable energy is gradually becoming an important member of the energy internet.Blockchain,as a decentralized distributed ledger with decentralization,traceability and tamper-proof features,is an importantway to achieve efficient consumption andmulti-party supply of new energy.In this article,we establish a blockchain-based mathematical model of multiple microgrids and microgrid aggregators’revenue,consider the degree of microgrid users’preference for electricity thus increasing users’reliance on the blockchainmarket,and apply the one-master-multiple-slave Stackelberg game theory to solve the energy dispatching strategy when each market entity pursues the maximum revenue.The simulation results show that the blockchain-based dynamic game of the multi-microgrid market can effectively increase the revenue of both microgrids and aggregators and improve the utilization of renewable energy.展开更多
In evolutionary games,most studies on finite populations have focused on a single updating mechanism.However,given the differences in individual cognition,individuals may change their strategies according to different...In evolutionary games,most studies on finite populations have focused on a single updating mechanism.However,given the differences in individual cognition,individuals may change their strategies according to different updating mechanisms.For this reason,we consider two different aspiration-driven updating mechanisms in structured populations:satisfied-stay unsatisfied shift(SSUS)and satisfied-cooperate unsatisfied defect(SCUD).To simulate the game player’s learning process,this paper improves the particle swarm optimization algorithm,which will be used to simulate the game player’s strategy selection,i.e.,population particle swarm optimization(PPSO)algorithms.We find that in the prisoner’s dilemma,the conditions that SSUS facilitates the evolution of cooperation do not enable cooperation to emerge.In contrast,SCUD conditions that promote the evolution of cooperation enable cooperation to emerge.In addition,the invasion of SCUD individuals helps promote cooperation among SSUS individuals.Simulated by the PPSO algorithm,the theoretical approximation results are found to be consistent with the trend of change in the simulation results.展开更多
In the realm of public goods game,punishment,as a potent tool,stands out for fostering cooperation.While it effectively addresses the first-order free-rider problem,the associated costs can be substantial.Punishers in...In the realm of public goods game,punishment,as a potent tool,stands out for fostering cooperation.While it effectively addresses the first-order free-rider problem,the associated costs can be substantial.Punishers incur expenses in imposing sanctions,while defectors face fines.Unfortunately,these monetary elements seemingly vanish into thin air,representing a loss to the system itself.However,by virtue of the redistribution of fines to cooperators and punishers,not only can we mitigate this loss,but the rewards for these cooperative individuals can be enhanced.Based upon this premise,this paper introduces a fine distribution mechanism to the traditional pool punishment model.Under identical parameter settings,by conducting a comparative experiment with the conventional punishment model,the paper aims to investigate the impact of fine distribution on the evolution of cooperation in spatial public goods game.The experimental results clearly demonstrate that,in instances where the punishment cost is prohibitively high,the cooperative strategies of the traditional pool punishment model may completely collapse.However,the model enriched with fine distribution manages to sustain a considerable number of cooperative strategies,thus highlighting its effectiveness in promoting and preserving cooperation,even in the face of substantial punishment cost.展开更多
This paper studies the evolutionary process of cooperative behavior in a public goods game model with heterogeneous investment strategies in square lattices.In the proposed model,players are divided into defectors,coo...This paper studies the evolutionary process of cooperative behavior in a public goods game model with heterogeneous investment strategies in square lattices.In the proposed model,players are divided into defectors,cooperators and discreet investors.Among these,defectors do not participate in investing,discreet investors make heterogeneous investments based on the investment behavior and cooperation value of their neighbors,and cooperators invest equally in each neighbor.In real life,heterogeneous investment is often accompanied by time or economic costs.The discreet investors in this paper pay a certain price to obtain their neighbors'investment behavior and cooperation value,which quantifies the time and economic costs of the heterogeneous investment process.The results of Monte Carlo simulation experiments in this study show that discreet investors can effectively resist the invasion of the defectors,form a stable cooperative group and expand the cooperative advantage in evolution.However,when discreet investors pay too high a price,they lose their strategic advantage.The results in this paper help us understand the role of heterogeneous investment in promoting and maintaining human social cooperation.展开更多
This paper presents a novel cooperative value iteration(VI)-based adaptive dynamic programming method for multi-player differential game models with a convergence proof.The players are divided into two groups in the l...This paper presents a novel cooperative value iteration(VI)-based adaptive dynamic programming method for multi-player differential game models with a convergence proof.The players are divided into two groups in the learning process and adapt their policies sequentially.Our method removes the dependence of admissible initial policies,which is one of the main drawbacks of the PI-based frameworks.Furthermore,this algorithm enables the players to adapt their control policies without full knowledge of others’ system parameters or control laws.The efficacy of our method is illustrated by three examples.展开更多
基金supported in part by the Strategic Priority Research Program of Chinese Academy of Sciences(XDA27030100)National Natural Science Foundation of China(72293575, 11832001)。
文摘The pursuit-evasion game models the strategic interaction among players, attracting attention in many realistic scenarios, such as missile guidance, unmanned aerial vehicles, and target defense. Existing studies mainly concentrate on the cooperative pursuit of multiple players in two-dimensional pursuit-evasion games. However, these approaches can hardly be applied to practical situations where players usually move in three-dimensional space with a three-degree-of-freedom control. In this paper,we make the first attempt to investigate the equilibrium strategy of the realistic pursuit-evasion game, in which the pursuer follows a three-degree-of-freedom control, and the evader moves freely. First, we describe the pursuer's three-degree-of-freedom control and the evader's relative coordinate. We then rigorously derive the equilibrium strategy by solving the retrogressive path equation according to the Hamilton-Jacobi-Bellman-Isaacs(HJBI) method, which divides the pursuit-evasion process into the navigation and acceleration phases. Besides, we analyze the maximum allowable speed for the pursuer to capture the evader successfully and provide the strategy with which the evader can escape when the pursuer's speed exceeds the threshold. We further conduct comparison tests with various unilateral deviations to verify that the proposed strategy forms a Nash equilibrium.
基金supported by the National Defense Science and Technology Innovation program(18-163-15-LZ-001-004-13).
文摘Current successes in artificial intelligence domain have revitalized interest in spacecraft pursuit-evasion game,which is an interception problem with a non-cooperative maneuvering target.The paper presents an automated machine learning(AutoML)based method to generate optimal trajectories in long-distance scenarios.Compared with conventional deep neural network(DNN)methods,the proposed method dramatically reduces the reliance on manual intervention and machine learning expertise.Firstly,based on differential game theory and costate normalization technique,the trajectory optimization problem is formulated under the assumption of continuous thrust.Secondly,the AutoML technique based on sequential model-based optimization(SMBO)framework is introduced to automate DNN design in deep learning process.If recommended DNN architecture exists,the tree-structured Parzen estimator(TPE)is used,otherwise the efficient neural architecture search(NAS)with network morphism is used.Thus,a novel trajectory optimization method with high computational efficiency is achieved.Finally,numerical results demonstrate the feasibility and efficiency of the proposed method.
基金supported by the National Defense Science and Techn ology Innovation(18-163-15-LZ-001-004-13)。
文摘With the development of space rendezvous and proximity operations(RPO)in recent years,the scenarios with noncooperative spacecraft are attracting the attention of more and more researchers.A method based on the costate normalization technique and deep neural networks is presented to generate the optimal guidance law for free-time orbital pursuit-evasion game.Firstly,the 24-dimensional problem given by differential game theory is transformed into a three-parameter optimization problem through the dimension-reduction method which guarantees the uniqueness of solution for the specific scenario.Secondly,a close-loop interactive mechanism involving feedback is introduced to deep neural networks for generating precise initial solution.Thus the optimal guidance law is obtained efficiently and stably with the application of optimization algorithm initialed by the deep neural networks.Finally,the results of the comparison with another two methods and Monte Carlo simulation demonstrate the efficiency and robustness of the proposed optimal guidance method.
文摘In this paper,the pursuit-evasion game with state and control constraints is solved to achieve the Nash equilibrium of both the pursuer and the evader with an iterative self-play technique.Under the condition where the Hamiltonian formed by means of Pontryagin’s maximum principle has the unique solution,it can be proven that the iterative control law converges to the Nash equilibrium solution.However,the strong nonlinearity of the ordinary differential equations formulated by Pontryagin’s maximum principle makes the control policy difficult to figured out.Moreover the system dynamics employed in this manuscript contains a high dimensional state vector with constraints.In practical applications,such as the control of aircraft,the provided overload is limited.Therefore,in this paper,we consider the optimal strategy of pursuit-evasion games with constant constraint on the control,while some state vectors are restricted by the function of the input.To address the challenges,the optimal control problems are transformed into nonlinear programming problems through the direct collocation method.Finally,two numerical cases of the aircraft pursuit-evasion scenario are given to demonstrate the effectiveness of the presented method to obtain the optimal control of both the pursuer and the evader.
基金supported,in part,by the National Natural Science Foundation of China(Nos.12272116 and 62088101)the Zhejiang Provincial Natural Science Foundation of China(Nos.LY22A020007 and LR20F030003)+1 种基金the Fundamental Research Funds for the Provincial Universities of Zhejiang,China(Nos.GK239909299001-014)the National Key Basic Research Strengthen Foundation of China(Nos.2021JCJQ-JJ-1183 and 2020-JCJQ-JJ-176)。
文摘This work is inspired by a stealth pursuit behavior called motion camouflage whereby a pursuer approaches an evader while the pursuer camouflages itself against a predetermined background.We formulate the spacecraft pursuit-evasion problem as a stealth pursuit strategy of motion camouflage,in which the pursuer tries to minimize a motion camouflage index defined in this paper.The Euler-Hill reference frame whose origin is set on the circular reference orbit is used to describe the dynamics.Based on the rule of motion camouflage,a guidance strategy in open-loop form to achieve motion camouflage index is derived in which the pursuer lies on the camouflage constraint line connecting the central spacecraft and evader.In order to dispose of the dependence on the evader acceleration in the open-loop guidance strategy,we further consider the motion camouflage pursuit problem within an infinite-horizon nonlinear quadratic differential game.The saddle point solution to the game is derived by using the state-dependent Riccati equation method,and the resulting closed-loop guidance strategy is effective in achieving motion camouflage.Simulations are performed to demonstrate the capabilities of the proposed guidance strategies for the pursuit–evasion game scenario.
文摘TheUAV pursuit-evasion problem focuses on the efficient tracking and capture of evading targets using unmanned aerial vehicles(UAVs),which is pivotal in public safety applications,particularly in scenarios involving intrusion monitoring and interception.To address the challenges of data acquisition,real-world deployment,and the limited intelligence of existing algorithms in UAV pursuit-evasion tasks,we propose an innovative swarm intelligencebased UAV pursuit-evasion control framework,namely“Boids Model-based DRL Approach for Pursuit and Escape”(Boids-PE),which synergizes the strengths of swarm intelligence from bio-inspired algorithms and deep reinforcement learning(DRL).The Boids model,which simulates collective behavior through three fundamental rules,separation,alignment,and cohesion,is adopted in our work.By integrating Boids model with the Apollonian Circles algorithm,significant improvements are achieved in capturing UAVs against simple evasion strategies.To further enhance decision-making precision,we incorporate a DRL algorithm to facilitate more accurate strategic planning.We also leverage self-play training to continuously optimize the performance of pursuit UAVs.During experimental evaluation,we meticulously designed both one-on-one and multi-to-one pursuit-evasion scenarios,customizing the state space,action space,and reward function models for each scenario.Extensive simulations,supported by the PyBullet physics engine,validate the effectiveness of our proposed method.The overall results demonstrate that Boids-PE significantly enhance the efficiency and reliability of UAV pursuit-evasion tasks,providing a practical and robust solution for the real-world application of UAV pursuit-evasion missions.
基金the National Natural Science Foundation of China(Grant Nos.11572345&11972044)the Program of National University of Defense Technology(Grant No.ZK18-03-07)。
文摘The orbital pursuit-evasion game is typically formulated as a complete-information game,which assumes the payoff functions of the two players are common knowledge.However,realistic pursuit-evasion games typically have incomplete information,in which the lack of payoff information limits the player’s ability to play optimally.To address this problem,this paper proposes a currently optimal escape strategy based on estimation for the evader.In this strategy,the currently optimal evasive controls are first derived based on the evader’s guess of the pursuer’s payoff weightings.Then an online parameter estimation method based on a modified strong tracking unscented Kalman filter is employed to modify the guess and update the strategy during the game.As the estimation becomes accurate,the currently optimal strategy gets closer to the actually optimal strategy.Simulation results show the proposed strategy can achieve optimal evasive controls progressively and the evader’s payoff of the strategy is lower than that of the zero-sum escape strategy.Meanwhile,the proposed strategy is also effective in the case where the pursuer changes its payoff function halfway during the game.
基金supported by Aeronautical Science Foundation of China(No.20160153002)National Natural Science Foundation of China(No.61933010)+1 种基金Aeronautical Science Foundation of China(No.20180753007)Natural Science Basic Research Plan in Shaanxi Province,China(No.2019JZ-08)。
文摘In practical combat scenario,the cooperative intercept strategies are often carefully designed,and it is challenging for the hypersonic vehicles to achieve successful evasion.Based on the analysis,it can be found that if several Successive Pursuers come from the Same Direction(SPSD)and flight with a proper spacing,the evasion difficulty may increase greatly.To address this problem,we focus on the evasion guidance strategy design for the Air-breathing Hypersonic Vehicles(AHVs)under the SPSD combat scenario.In order to avoid the induced influence on the scramjet,altitude and speed of the vehicle,the lateral maneuver and evasion are employed.To guarantee the remnant maneuver ability,the concept of specified miss distance is introduced and utilized to generate the guidance command for the AHV.In the framework of constrained optimal control,the analytical expression of the evasion command is derived,and the constraints of the overload can be ensured to be never violated.In fact,by analyzing the spacing of the pursers,it can be classified whether the cooperative pursuit is formed.For the coordination-unformed multiple pursers,the evasion can be achieved lightly by the proposed strategy.If the coordination is formed,the proposed method will generate a large reverse direction maneuver,and the successful evasion can be maintained as a result.The performance of the proposed algorithms is tested in numerical simulations.
基金supported in part by the National Natural Science Foundation of China (No.62271253,61901523,62001381)Fundamental Research Funds for the Central Universities (No.NS2023018)+2 种基金the National Aerospace Science Foundation of China under Grant 2023Z021052002the open research fund of National Mobile Communications Research Laboratory,Southeast University (No.2023D09)Postgraduate Research & Practice Innovation Program of NUAA (No.xcxjh20220402)。
文摘To improve the anti-jamming and interference mitigation ability of the UAV-aided communication systems, this paper investigates the channel selection optimization problem in face of both internal mutual interference and external malicious jamming. A cooperative anti-jamming and interference mitigation method based on local altruistic is proposed to optimize UAVs’ channel selection. Specifically, a Stackelberg game is modeled to formulate the confrontation relationship between UAVs and the jammer. A local altruistic game is modeled with each UAV considering the utilities of both itself and other UAVs. A distributed cooperative anti-jamming and interference mitigation algorithm is proposed to obtain the Stackelberg equilibrium. Finally, the convergence of the proposed algorithm and the impact of the transmission power on the system loss value are analyzed, and the anti-jamming performance of the proposed algorithm can be improved by around 64% compared with the existing algorithms.
基金sponsored by the National Key R&D Program of China(No.2018YFB2100400)the National Natural Science Foundation of China(No.62002077,61872100)+4 种基金the Major Research Plan of the National Natural Science Foundation of China(92167203)the Guangdong Basic and Applied Basic Research Foundation(No.2020A1515110385)the China Postdoctoral Science Foundation(No.2022M710860)the Zhejiang Lab(No.2020NF0AB01)Guangzhou Science and Technology Plan Project(202102010440).
文摘Benefiting from the development of Federated Learning(FL)and distributed communication systems,large-scale intelligent applications become possible.Distributed devices not only provide adequate training data,but also cause privacy leakage and energy consumption.How to optimize the energy consumption in distributed communication systems,while ensuring the privacy of users and model accuracy,has become an urgent challenge.In this paper,we define the FL as a 3-layer architecture including users,agents and server.In order to find a balance among model training accuracy,privacy-preserving effect,and energy consumption,we design the training process of FL as game models.We use an extensive game tree to analyze the key elements that influence the players’decisions in the single game,and then find the incentive mechanism that meet the social norms through the repeated game.The experimental results show that the Nash equilibrium we obtained satisfies the laws of reality,and the proposed incentive mechanism can also promote users to submit high-quality data in FL.Following the multiple rounds of play,the incentive mechanism can help all players find the optimal strategies for energy,privacy,and accuracy of FL in distributed communication systems.
基金supported by the National Natural Science Foundation of China(Nos.51977113,62293500,62293501 and 62293505).
文摘Malicious attacks against data are unavoidable in the interconnected,open and shared Energy Internet(EI),Intrusion tolerant techniques are critical to the data security of EI.Existing intrusion tolerant techniques suffered from problems such as low adaptability,policy lag,and difficulty in determining the degree of tolerance.To address these issues,we propose a novel adaptive intrusion tolerance model based on game theory that enjoys two-fold ideas:(1)it constructs an improved replica of the intrusion tolerance model of the dynamic equation evolution game to induce incentive weights;and (2)it combines a tournament competition model with incentive weights to obtain optimal strategies for each stage of the game process.Extensive experiments are conducted in the IEEE 39-bus system,whose results demonstrate the feasibility of the incentive weights,confirm the proposed strategy strengthens the system’s ability to tolerate aggression,and improves the dynamic adaptability and response efficiency of the aggression-tolerant system in the case of limited resources.
文摘1.Introduction In August 2024,over 4400 Paralympic athletes will gather in Paris for the Paralympic Summer Games—the pinnacle of every Paralympian’s(Para athletes competing at the Paralympic Games)career to showcase their ability and skills.Their training,preparation,and effort in the years leading up to the Games are unparalleled.To achieve success,Paralympians specifically rely on a medical support team to achieve their goals.So,what is required of the medical support team to prepare Paralympians to get ready,set,and go to Paris 2024?
文摘Objective: To study the problematic use of video games among secondary school students in the city of Parakou in 2023. Methods: Descriptive cross-sectional study conducted in the commune of Parakou from December 2022 to July 2023. The study population consisted of students regularly enrolled in public and private secondary schools in the city of Parakou for the 2022-2023 academic year. A two-stage non-proportional stratified sampling technique combined with simple random sampling was adopted. The Problem Video Game Playing (PVP) scale was used to assess problem gambling in the study population, while anxiety and depression were assessed using the Hospital Anxiety and Depression Scale (HADS). Results: A total of 1030 students were included. The mean age of the pupils surveyed was 15.06 ± 2.68 years, with extremes of 10 and 28 years. The [13 - 18] age group was the most represented, with a proportion of 59.6% (614) in the general population. Females predominated, at 52.8% (544), with a sex ratio of 0.89. The prevalence of problematic video game use was 24.9%, measured using the Video Game Playing scale. Associated factors were male gender (p = 0.005), pocket money under 10,000 cfa (p = 0.001) and between 20,000 - 90,000 cfa (p = 0.030), addictive family behavior (p < 0.001), monogamous family (p = 0.023), good relationship with father (p = 0.020), organization of video game competitions (p = 0.001) and definite anxiety (p Conclusion: Substance-free addiction is struggling to attract the attention it deserves, as it did in its infancy everywhere else. This study complements existing data and serves as a reminder of the need to focus on this group of addictions, whose problematic use of video games remains the most frequent due to its accessibility and social tolerance. Preventive action combined with curative measures remains the most effective means of combating the problem at national level.
文摘Given a graph g=( V,A ) , we define a space of subgraphs M with the binary operation of union and the unique decomposition property into blocks. This space allows us to discuss a notion of minimal subgraphs (minimal coalitions) that are of interest for the game. Additionally, a partition of the game is defined in terms of the gain of each block, and subsequently, a solution to the game is defined based on distributing to each player (node and edge) present in each block a payment proportional to their contribution to the coalition.
基金the National Natural Science Foun-dation of China(Grant No.71961003).
文摘In public goods games, punishments and rewards have been shown to be effective mechanisms for maintaining individualcooperation. However, punishments and rewards are costly to incentivize cooperation. Therefore, the generation ofcostly penalties and rewards has been a complex problem in promoting the development of cooperation. In real society,specialized institutions exist to punish evil people or reward good people by collecting taxes. We propose a strong altruisticpunishment or reward strategy in the public goods game through this phenomenon. Through theoretical analysis and numericalcalculation, we can get that tax-based strong altruistic punishment (reward) has more evolutionary advantages thantraditional strong altruistic punishment (reward) in maintaining cooperation and tax-based strong altruistic reward leads toa higher level of cooperation than tax-based strong altruistic punishment.
基金This research was funded by the NSFC under Grant No.61803279in part by the Qing Lan Project of Jiangsu,in part by the China Postdoctoral Science Foundation under Grant Nos.2020M671596 and 2021M692369+3 种基金in part by the Suzhou Science and Technology Development Plan Project(Key Industry Technology Innovation)under Grant No.SYG202114in part by the Open Project Funding from Anhui Province Key Laboratory of Intelligent Building and Building Energy Saving,Anhui Jianzhu University,under Grant No.IBES2021KF08in part by the Postgraduate Research&Practice Innovation Program of Jiangsu Province under Grant No.KYCX23_3320in part by the Postgraduate Research&Practice Innovation Program of Jiangsu Province under Grant No.SJCX22_1585.
文摘As the current global environment is deteriorating,distributed renewable energy is gradually becoming an important member of the energy internet.Blockchain,as a decentralized distributed ledger with decentralization,traceability and tamper-proof features,is an importantway to achieve efficient consumption andmulti-party supply of new energy.In this article,we establish a blockchain-based mathematical model of multiple microgrids and microgrid aggregators’revenue,consider the degree of microgrid users’preference for electricity thus increasing users’reliance on the blockchainmarket,and apply the one-master-multiple-slave Stackelberg game theory to solve the energy dispatching strategy when each market entity pursues the maximum revenue.The simulation results show that the blockchain-based dynamic game of the multi-microgrid market can effectively increase the revenue of both microgrids and aggregators and improve the utilization of renewable energy.
基金Project supported by the Doctoral Foundation Project of Guizhou University(Grant No.(2019)49)the National Natural Science Foundation of China(Grant No.71961003)the Science and Technology Program of Guizhou Province(Grant No.7223)。
文摘In evolutionary games,most studies on finite populations have focused on a single updating mechanism.However,given the differences in individual cognition,individuals may change their strategies according to different updating mechanisms.For this reason,we consider two different aspiration-driven updating mechanisms in structured populations:satisfied-stay unsatisfied shift(SSUS)and satisfied-cooperate unsatisfied defect(SCUD).To simulate the game player’s learning process,this paper improves the particle swarm optimization algorithm,which will be used to simulate the game player’s strategy selection,i.e.,population particle swarm optimization(PPSO)algorithms.We find that in the prisoner’s dilemma,the conditions that SSUS facilitates the evolution of cooperation do not enable cooperation to emerge.In contrast,SCUD conditions that promote the evolution of cooperation enable cooperation to emerge.In addition,the invasion of SCUD individuals helps promote cooperation among SSUS individuals.Simulated by the PPSO algorithm,the theoretical approximation results are found to be consistent with the trend of change in the simulation results.
基金the Open Foundation of Key Lab-oratory of Software Engineering of Yunnan Province(Grant Nos.2020SE308 and 2020SE309).
文摘In the realm of public goods game,punishment,as a potent tool,stands out for fostering cooperation.While it effectively addresses the first-order free-rider problem,the associated costs can be substantial.Punishers incur expenses in imposing sanctions,while defectors face fines.Unfortunately,these monetary elements seemingly vanish into thin air,representing a loss to the system itself.However,by virtue of the redistribution of fines to cooperators and punishers,not only can we mitigate this loss,but the rewards for these cooperative individuals can be enhanced.Based upon this premise,this paper introduces a fine distribution mechanism to the traditional pool punishment model.Under identical parameter settings,by conducting a comparative experiment with the conventional punishment model,the paper aims to investigate the impact of fine distribution on the evolution of cooperation in spatial public goods game.The experimental results clearly demonstrate that,in instances where the punishment cost is prohibitively high,the cooperative strategies of the traditional pool punishment model may completely collapse.However,the model enriched with fine distribution manages to sustain a considerable number of cooperative strategies,thus highlighting its effectiveness in promoting and preserving cooperation,even in the face of substantial punishment cost.
基金Project supported by the Open Foundation of Key Laboratory of Software Engineering of Yunnan Province(Grant Nos.2020SE308 and 2020SE309).
文摘This paper studies the evolutionary process of cooperative behavior in a public goods game model with heterogeneous investment strategies in square lattices.In the proposed model,players are divided into defectors,cooperators and discreet investors.Among these,defectors do not participate in investing,discreet investors make heterogeneous investments based on the investment behavior and cooperation value of their neighbors,and cooperators invest equally in each neighbor.In real life,heterogeneous investment is often accompanied by time or economic costs.The discreet investors in this paper pay a certain price to obtain their neighbors'investment behavior and cooperation value,which quantifies the time and economic costs of the heterogeneous investment process.The results of Monte Carlo simulation experiments in this study show that discreet investors can effectively resist the invasion of the defectors,form a stable cooperative group and expand the cooperative advantage in evolution.However,when discreet investors pay too high a price,they lose their strategic advantage.The results in this paper help us understand the role of heterogeneous investment in promoting and maintaining human social cooperation.
基金supported by the Industry-University-Research Cooperation Fund Project of the Eighth Research Institute of China Aerospace Science and Technology Corporation (USCAST2022-11)Aeronautical Science Foundation of China (20220001057001)。
文摘This paper presents a novel cooperative value iteration(VI)-based adaptive dynamic programming method for multi-player differential game models with a convergence proof.The players are divided into two groups in the learning process and adapt their policies sequentially.Our method removes the dependence of admissible initial policies,which is one of the main drawbacks of the PI-based frameworks.Furthermore,this algorithm enables the players to adapt their control policies without full knowledge of others’ system parameters or control laws.The efficacy of our method is illustrated by three examples.