论文标题
ARLO:自动加固学习框架
ARLO: A Framework for Automated Reinforcement Learning
论文作者
论文摘要
自动化增强学习(AUTORL)是一个相对较新的研究领域,正在越来越关注。 Autorl的目的是通过减轻其一些主要挑战,包括数据收集,算法选择和超参数调整,从而减轻更广泛的公众的增强学习(RL)技术。在这项工作中,我们提出了一个一般且灵活的框架,即Arlo:自动增强学习优化器,以构建自动化的自动化管道。基于此,我们提出了一条离线管道和在线RL的管道,讨论了组件,交互并突出两种设置之间的区别。此外,我们提供了作为开源库发布的此类管道的Python实施。我们的实施已在说明性的LQG域和经典的Mujoco环境上进行了测试,显示了需要有限的人类干预的竞争性能的能力。我们还在现实的大坝环境上展示了完整的管道,自动执行功能选择和模型生成任务。
Automated Reinforcement Learning (AutoRL) is a relatively new area of research that is gaining increasing attention. The objective of AutoRL consists in easing the employment of Reinforcement Learning (RL) techniques for the broader public by alleviating some of its main challenges, including data collection, algorithm selection, and hyper-parameter tuning. In this work, we propose a general and flexible framework, namely ARLO: Automated Reinforcement Learning Optimizer, to construct automated pipelines for AutoRL. Based on this, we propose a pipeline for offline and one for online RL, discussing the components, interaction, and highlighting the difference between the two settings. Furthermore, we provide a Python implementation of such pipelines, released as an open-source library. Our implementation has been tested on an illustrative LQG domain and on classic MuJoCo environments, showing the ability to reach competitive performances requiring limited human intervention. We also showcase the full pipeline on a realistic dam environment, automatically performing the feature selection and the model generation tasks.