论文标题
模态自适应面部识别的领域私人和不可知论功能
Domain Private and Agnostic Feature for Modality Adaptive Face Recognition
论文作者
论文摘要
由于较大的方式差异和跨模式样本不足,因此异质的面部识别是一项具有挑战性的任务。大多数现有的作品都集中在判别特征转换,度量学习和跨模式的面部合成上。但是,跨模式面孔总是与域(模式)和身份信息相结合的事实很少受到关注。因此,如何学习和利用域特征和域 - 不合时宜的特征以进行模态自适应面部识别是这项工作的重点。具体而言,本文提出了一个功能聚合网络(FAN),其中包括分离的表示模块(DRM),特征融合模块(FFM)和自适应惩罚度量指标(APM)学习会话。首先,在DRM中,两个子网,即域 - 私有网络和域 - 不合命替网络的专门设计,分别用于学习模态特征和身份特征。其次,在FFM中,身份特征与域特征融合在一起,以实现跨模式双向身份特征转换,在很大程度上,这进一步删除了模态信息和身份信息。第三,考虑到跨模式数据集中存在简易和硬对之间的分布不平衡,这增加了模型偏差的风险,因此在我们的粉丝中提出了通过自适应硬对惩罚来保存指导的指标学习的身份。所提出的APM还保证了跨模式的类内部紧凑性和阶层间分离。基准跨模式面部数据集的广泛实验表明,我们的风扇优于SOTA方法。
Heterogeneous face recognition is a challenging task due to the large modality discrepancy and insufficient cross-modal samples. Most existing works focus on discriminative feature transformation, metric learning and cross-modal face synthesis. However, the fact that cross-modal faces are always coupled by domain (modality) and identity information has received little attention. Therefore, how to learn and utilize the domain-private feature and domain-agnostic feature for modality adaptive face recognition is the focus of this work. Specifically, this paper proposes a Feature Aggregation Network (FAN), which includes disentangled representation module (DRM), feature fusion module (FFM) and adaptive penalty metric (APM) learning session. First, in DRM, two subnetworks, i.e. domain-private network and domain-agnostic network are specially designed for learning modality features and identity features, respectively. Second, in FFM, the identity features are fused with domain features to achieve cross-modal bi-directional identity feature transformation, which, to a large extent, further disentangles the modality information and identity information. Third, considering that the distribution imbalance between easy and hard pairs exists in cross-modal datasets, which increases the risk of model bias, the identity preserving guided metric learning with adaptive hard pairs penalization is proposed in our FAN. The proposed APM also guarantees the cross-modality intra-class compactness and inter-class separation. Extensive experiments on benchmark cross-modal face datasets show that our FAN outperforms SOTA methods.