论文标题
与未观察到的混杂因素以公正的建议重新考虑学习目标
Reconsidering Learning Objectives in Unbiased Recommendation with Unobserved Confounders
论文作者
论文摘要
这项工作研究了从偏见反馈中学习无偏见算法的问题,以提出建议。我们从新颖的分发转变角度解决了这个问题。无偏见的最新作品通过各种技术(例如重新加权,多任务学习和元学习)提出了最新的技术。尽管取得了经验的成功,但大多数人都缺乏理论保证,从而在理论和最近的算法之间形成了不可忽略的差距。在本文中,我们提出了一个理论上的理解,以了解为什么现有的无偏学习目标有助于无偏见的建议。我们建立了无偏见的建议和分配变化之间的密切联系,这表明现有的无偏学习目标隐含地偏向偏见的培训和无偏见的测试分布。基于这种联系,我们为现有的无偏学习方法开发了两个概括范围,并分析了他们的学习行为。此外,由于分配变化,我们进一步提出了一个原则上的框架,对抗性自我训练(AST),以无偏见的建议。对现实世界和半合成数据集的广泛实验证明了AST的有效性。
This work studies the problem of learning unbiased algorithms from biased feedback for recommendation. We address this problem from a novel distribution shift perspective. Recent works in unbiased recommendation have advanced the state-of-the-art with various techniques such as re-weighting, multi-task learning, and meta-learning. Despite their empirical successes, most of them lack theoretical guarantees, forming non-negligible gaps between theories and recent algorithms. In this paper, we propose a theoretical understanding of why existing unbiased learning objectives work for unbiased recommendation. We establish a close connection between unbiased recommendation and distribution shift, which shows that existing unbiased learning objectives implicitly align biased training and unbiased test distributions. Built upon this connection, we develop two generalization bounds for existing unbiased learning methods and analyze their learning behavior. Besides, as a result of the distribution shift, we further propose a principled framework, Adversarial Self-Training (AST), for unbiased recommendation. Extensive experiments on real-world and semi-synthetic datasets demonstrate the effectiveness of AST.