摘要
Dynamic channel assignment(DCA)plays a key role in extending vehicular ad-hoc network capacity and mitigating congestion.However,channel assignment under vehicular direct communication scenarios faces mutual influence of large-scale nodes,the lack of centralized coordination,unknown global state information,and other challenges.To solve this problem,a multiagent reinforcement learning(RL)based cooperative DCA(RLCDCA)mechanism is proposed.Specifically,each vehicular node can successfully learn the proper strategies of channel selection and backoff adaptation from the real-time channel state information(CSI)using two cooperative RL models.In addition,neural networks are constructed as nonlinear Q-function approximators,which facilitates the mapping of the continuously sensed input to the mixed policy output.Nodes are driven to locally share and incorporate their individual rewards such that they can optimize their policies in a distributed collaborative manner.Simulation results show that the proposed multiagent RL-CDCA can better reduce the one-hop packet delay by no less than 73.73%,improve the packet delivery ratio by no less than 12.66%on average in a highly dense situation,and improve the fairness of the global network resource allocation.
基金
Project supported by the National Natural Science Foundation of China(Nos.61672082 and 61822101)
the Beijing Municipal Natural Science Foundation,China(No.4181002)
the Beihang University Innovation and Practice Fund for Graduate,China(No.YCSJ-02-2018-05)。