期刊文献+

基于顶点冲突学习的最大公共子图算法

Maximum common induced subgraph algorithm based on vertex conflict learning
下载PDF
导出
摘要 针对最大公共子图(MCS)的传统分支策略依赖于图的静态属性,缺少学习历史搜索信息的问题,提出了基于顶点冲突学习的分支策略。首先,把上界的减少值作为分支点完成匹配动作的奖励;其次,由于当最优解被更新时,得到的最优解是分支点不断推理产生的结果,因此给予在完整的搜索路径上的分支点适当的奖励,从而强化这些顶点对搜索的积极作用;最后,设计了匹配动作的价值函数,并选择具有最大累计奖励的顶点作为新的分支点。在McSplit算法基础上,提出了糅合新分支策略的McSplitRLR算法。实验结果表明,除去均可以被所有对比算法在10 s之内解决的简单算例,在相同机器和求解限制时间条件下,相较当前先进的算法McSplit、McSplitSBS,McSplitRLR分别多解决了109、33个困难算例,求解率分别提高了5.6%、1.6%。 The traditional branching strategies of Maximum Common induced Subgraph(MCS)problem rely on the static properties of graphs and lack learning information about historical searches.In order to solve these problems,a branching strategy based on vertex conflict learning was proposed.Firstly,the reduction value of the upper bound was used as the reward to the branch node for completing a matching action.Secondly,when the optimal solution was updated,the optimal solution obtained actually was the result of continuous inference of the branch nodes.Therefore,the appropriate rewards were given to the branch nodes on the complete search path to strengthen the positive effect of these vertices on search.Finally,the value function of matching action was designed,and the vertices with the maximum cumulative rewards would be selected as new branch nodes.On the basis of Maximum common induced subgraph Split(McSplit)algorithm,an improved McSplit Reinforcement Learning and Routing(McSplitRLR)algorithm combined with the new branching strategy was completed.Experimental results show that,with the same computer and solution time limit,excluding the simple instances solved by all comparison algorithms within 10 seconds,compared to the state-of-the-art algorithms of McSplit and McSplit Solution-Biased Search(McSplitSBS),McSplitRLR solves 109 and 33 more hard instances respectively,and the solution rate increases by 5.6%and 1.6%respectively.
作者 王宇 刘燕丽 陈劭武 WANG Yu;LIU Yanli;CHEN Shaowu(College of Science,Wuhan University of Science and Technology,Wuhan Hubei 430081,China;Hubei Province Key Laboratory of Systems Science in Metallurgical Process(Wuhan University of Science and Technology),Wuhan Hubei 430081,China)
出处 《计算机应用》 CSCD 北大核心 2021年第6期1756-1760,共5页 journal of Computer Applications
基金 湖北省大学生创新训练项目(S201910488044) 冶金工业过程系统科学湖北重点实验室开放基金资助项目(Y201716)。
关键词 组合优化问题 NP-HARD问题 强化学习 算法设计 最大公共子图 combinatorial optimization problem Non-deterministic Polynomial Hard(NP-Hard)problem reinforcement learning algorithm design Maximum Common induced Subgraph(MCS)
  • 相关文献

参考文献3

二级参考文献63

  • 1杨洋,陈小平.动态不确定环境下的决策:一种分层决策模型[J].计算机科学,2005,32(1):151-154. 被引量:1
  • 2苏畅,高阳,陈世福,陈兆乾.基于SMDP环境的自主生成options算法的研究[J].模式识别与人工智能,2005,18(6):679-684. 被引量:9
  • 3秦志斌,钱徽,朱淼良.自主移动机器人混合式体系结构的一种Multi-agent实现方法[J].机器人,2006,28(5):478-482. 被引量:8
  • 4原魁,李园,房立新.多移动机器人系统研究发展近况[J].自动化学报,2007,33(8):785-794. 被引量:73
  • 5AL-BATAH M S,MATISA N A,ZAMLI K Z,et al.Modified recursive least squares algorithm to train the hybrid multilayered perceptron (HMLP) network[J].Applied Soft Computing,2010,10(1):236-244.
  • 6BOWLING M.Multi agent learning in the presence of agents with limi-tations[R].Pittsburgh:Carnegie Mellon University,2003.
  • 7KYUN Y,OH S-Y.Hybrid control for autonomous mobile robotnavigation using neural network based behavior modules and environment classification[J].Autonomous Robots,2003,15(2):193-206.
  • 8ARAI S,SYCARA K.Multi-agent reinforcement learning for planning and conflict resolution in a dynamic domain[C] //Proc of the 4th International Conference on Autonomous agents.2000:104-105.
  • 9VRANCY P,VERBEEK K,NOWE A.Decetralized learning in Markov games[J].IEEE Trans on Systems,Man and Cyberne-tics Part B:Cybernetics,2008,38(4):976-981.
  • 10LUCIAN B,ROBERT B,BART D S.A comprehension survey of multiagent reinforcement learning[J].IEEE Trans on Systems,Man and Cybernetics Part C:Applications and Reviews,2008,68(2):156-172.

共引文献64

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部