...
首页> 外文期刊>Expert Systems with Application >Reactive Search strategies using Reinforcement Learning, local search algorithms and Variable Neighborhood Search
【24h】

Reactive Search strategies using Reinforcement Learning, local search algorithms and Variable Neighborhood Search

机译:使用强化学习,本地搜索算法和可变邻域搜索的被动搜索策略

获取原文
获取原文并翻译 | 示例
           

摘要

Optimization techniques known as metaheuristics have been applied successfully to solve different problems, in which their development is characterized by the appropriate selection of parameters (values) for its execution. Where the adjustment of a parameter is required, this parameter will be tested until viable results are obtained. Normally, such adjustments are made by the developer deploying the metaheuristic. The quality of the results of a test instance [The term instance is used to refer to the assignment of values to the input variables of a problem.] will not be transferred to the instances that were not tested yet and its feedback may require a slow process of "trial and error" where the algorithm has to be adjusted for a specific application. Within this context of metaheuristics the Reactive Search emerged defending the integration of machine learning within heuristic searches for solving complex optimization problems. Based in the integration that the Reactive Search proposes between machine learning and metaheuristics, emerged the idea of putting Reinforcement Learning, more specifically the Q-learning algorithm with a reactive behavior, to select which local search is the most appropriate in a given time of a search, to succeed another local search that can not improve the current solution in the VNS metaheuristic. In this work we propose a reactive implementation using Reinforcement Learning for the self-tuning of the implemented algorithm, applied to the Symmetric Travelling Salesman Problem.
机译:被称为元启发式优化技术的优化技术已成功应用于解决不同的问题,在这些技术的发展中,其特征在于对其执行进行适当选择的参数(值)的选择。需要调整参数的地方,将测试该参数,直到获得可行的结果为止。通常,此类调整是由部署元启发式方法的开发人员进行的。测试实例结果的质量[术语“实例用于指代对问题的输入变量的值分配。”将不会转移到尚未测试的实例,并且其反馈可能需要缓慢的时间。 “试验和错误”的过程,其中必须针对特定应用调整算法。在这种元启发式方法的背景下,反应式搜索应运而生,捍卫了机器学习在启发式搜索中的集成,以解决复杂的优化问题。在反应式搜索在机器学习和元启发式方法之间提出的整合基础上,出现了将强化学习(更具体地讲是具有反应性行为的Q学习算法)用于选择哪个局部搜索在特定时间段内最合适的想法。搜索,以成功完成另一项无法改善VNS元启发式方法当前解决方案的本地搜索。在这项工作中,我们提出了一种使用强化学习的反应式实施方案,以实现所实施算法的自调整,并将其应用于对称旅行商问题。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号