论文标题
通过推出风险(REX)的分布概括(REX)
Out-of-Distribution Generalization via Risk Extrapolation (REx)
论文作者
论文摘要
分配转移是将机器学习预测系统从实验室转移到现实世界的主要障碍之一。为了解决这个问题,我们假设跨训练域的变化代表了我们在测试时可能遇到的变化,但在测试时间的变化也可能更为极端。特别是,我们表明,减少培训域中的风险差异可以降低模型对广泛的极端分布变化的敏感性,包括在挑战性的环境中既包含因果关系和抗果元素。我们激励这种方法,风险外推(REX),是对外推域(MM-REX)扰动集的一种强大优化形式,并提出了对培训风险差异(V-Rex)的惩罚,作为一个简单的变体。我们证明,雷克斯的变体可以恢复目标的因果机制,同时还为输入分布的变化提供了一些稳健性(“协方差偏移”)。通过适当地交易因果诱发的分布变化和协变量转移,Rex能够超过替代方法,例如在这些类型的转移共发生的情况下,诸如不变风险最小化。
Distributional shift is one of the major obstacles when transferring machine learning prediction systems from the lab to the real world. To tackle this problem, we assume that variation across training domains is representative of the variation we might encounter at test time, but also that shifts at test time may be more extreme in magnitude. In particular, we show that reducing differences in risk across training domains can reduce a model's sensitivity to a wide range of extreme distributional shifts, including the challenging setting where the input contains both causal and anti-causal elements. We motivate this approach, Risk Extrapolation (REx), as a form of robust optimization over a perturbation set of extrapolated domains (MM-REx), and propose a penalty on the variance of training risks (V-REx) as a simpler variant. We prove that variants of REx can recover the causal mechanisms of the targets, while also providing some robustness to changes in the input distribution ("covariate shift"). By appropriately trading-off robustness to causally induced distributional shifts and covariate shift, REx is able to outperform alternative methods such as Invariant Risk Minimization in situations where these types of shift co-occur.