论文标题
关于团体公平与属性隐私的对齐
On the Alignment of Group Fairness with Attribute Privacy
论文作者
论文摘要
团体公平和隐私是设计值得信赖的机器学习模型的基本方面。先前的研究强调了群体公平与不同的隐私概念之间的冲突。我们是第一个证明团体公平性与黑框设置中属性隐私的特定隐私概念的一致性的人。通过对属性推理攻击(AIA)的阻力量化的属性隐私需要在目标模型的输出预测中不可区分。团体公平保证了这一点,从而减轻AIA并实现属性隐私。为了证明这一点,我们首先介绍了Adaptaia,这是对现有AIA的增强,该AIA是针对现实世界中具有敏感属性类不平衡的现实数据集量身定制的。通过理论和广泛的经验分析,我们证明了两种标准组公平算法(即对Adaptaia的对抗性偏见和疏远梯度下降)的功效。此外,由于使用群体公平会导致属性隐私,因此它是针对AIA的防御,目前缺乏。总体而言,我们表明,除了已经存在的模型实用程序的权衡权衡外,群体的公平性与属性隐私一致。
Group fairness and privacy are fundamental aspects in designing trustworthy machine learning models. Previous research has highlighted conflicts between group fairness and different privacy notions. We are the first to demonstrate the alignment of group fairness with the specific privacy notion of attribute privacy in a blackbox setting. Attribute privacy, quantified by the resistance to attribute inference attacks (AIAs), requires indistinguishability in the target model's output predictions. Group fairness guarantees this thereby mitigating AIAs and achieving attribute privacy. To demonstrate this, we first introduce AdaptAIA, an enhancement of existing AIAs, tailored for real-world datasets with class imbalances in sensitive attributes. Through theoretical and extensive empirical analyses, we demonstrate the efficacy of two standard group fairness algorithms (i.e., adversarial debiasing and exponentiated gradient descent) against AdaptAIA. Additionally, since using group fairness results in attribute privacy, it acts as a defense against AIAs, which is currently lacking. Overall, we show that group fairness aligns with attribute privacy at no additional cost other than the already existing trade-off with model utility.