论文标题

在线控制自适应大型邻里搜索的搜索

Online Control of Adaptive Large Neighborhood Search using Deep Reinforcement Learning

论文作者

Reijnen, Robbert, Zhang, Yingqian, Lau, Hoong Chuin, Bukhsh, Zaharah

论文摘要

自适应大型邻里搜索(ALNS)算法在解决组合优化问题(COPS)方面表现出了很大的成功。但是,ALN的性能依赖于其选择和接受参数的正确配置,这是一项复杂且资源密集的任务。为了解决这个问题,我们介绍了一种基于深的增强学习(DRL)的方法,称为DR-ALNS,该方法选择了操作员,调整参数并控制整个搜索过程中的接受标准。提出的方法旨在根据搜索的状态学习为下一个迭代配置ALN,以便为给定优化问题提供更有效的解决方案。我们在IJCAI竞争中提出的随机重量和时间窗口的定向问题评估了提出的方法。结果表明,我们的方法优于贝叶斯优化的Alns的香草Alns和两种最先进的DRL方法,这些方法是竞争的获胜方法,以较少的训练观察结果来实现这一目标。此外,我们证明了所提出的DR-ALNS方法的几种良好特性:它很容易适应解决不同的路由问题,其学识渊博的策略在各种实例大小上都始终如一地表现出色,并且这些策略可以直接应用于不同的问题变体。

The Adaptive Large Neighborhood Search (ALNS) algorithm has shown considerable success in solving combinatorial optimization problems (COPs). Nonetheless, the performance of ALNS relies on the proper configuration of its selection and acceptance parameters, which is known to be a complex and resource-intensive task. To address this, we introduce a Deep Reinforcement Learning (DRL) based approach called DR-ALNS that selects operators, adjusts parameters, and controls the acceptance criterion throughout the search. The proposed method aims to learn, based on the state of the search, to configure ALNS for the next iteration to yield more effective solutions for the given optimization problem. We evaluate the proposed method on an orienteering problem with stochastic weights and time windows, as presented in an IJCAI competition. The results show that our approach outperforms vanilla ALNS, ALNS tuned with Bayesian optimization, and two state-of-the-art DRL approaches that were the winning methods of the competition, achieving this with significantly fewer training observations. Furthermore, we demonstrate several good properties of the proposed DR-ALNS method: it is easily adapted to solve different routing problems, its learned policies perform consistently well across various instance sizes, and these policies can be directly applied to different problem variants.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源