TY - JOUR
T1 - Fairness-aware training of face attribute classifiers via adversarial robustness
AU - Zeng, Huimin
AU - Yue, Zhenrui
AU - Kou, Ziyi
AU - Zhang, Yang
AU - Shang, Lanyu
AU - Wang, Dong
N1 - Publisher Copyright:
© 2023 Elsevier B.V.
PY - 2023/3/15
Y1 - 2023/3/15
N2 - Developing fair deep learning models for identity-sensitive applications (e.g., face attribute recognition) has gained increasing attention from the research community. Indeed, it has been observed that deep models can easily overfit to the bias of the training set, resulting in discriminative performance against certain demographic groups during test time. Motivated by the observation that a biased classifier could result in different adversarial robustness among training samples in different demographic groups (robustness bias), we argue that such adversarial robustness information of individual training samples could imply whether the training data distribution is fair among different demographic groups. In other words, under a fair classifier, the training samples from different demographic groups are expected to show similar or comparable adversarial robustness. Therefore, in this work, we propose to re-weight the training loss of individual training samples using their adversarial robustness, and provide the fairness-awareness in the training process. Extensive experimental results on CelebA dataset show that the face attribute classifiers could learn significantly greater demographic fairness under our proposed training objective and outperform other state-of-the-art re-weighting fairness algorithms on different face recognition applications. Moreover, our proposed method also reduces the non-trivial robustness bias among different demographic groups, preventing the under-represented demographic groups from higher adversarial threats.
AB - Developing fair deep learning models for identity-sensitive applications (e.g., face attribute recognition) has gained increasing attention from the research community. Indeed, it has been observed that deep models can easily overfit to the bias of the training set, resulting in discriminative performance against certain demographic groups during test time. Motivated by the observation that a biased classifier could result in different adversarial robustness among training samples in different demographic groups (robustness bias), we argue that such adversarial robustness information of individual training samples could imply whether the training data distribution is fair among different demographic groups. In other words, under a fair classifier, the training samples from different demographic groups are expected to show similar or comparable adversarial robustness. Therefore, in this work, we propose to re-weight the training loss of individual training samples using their adversarial robustness, and provide the fairness-awareness in the training process. Extensive experimental results on CelebA dataset show that the face attribute classifiers could learn significantly greater demographic fairness under our proposed training objective and outperform other state-of-the-art re-weighting fairness algorithms on different face recognition applications. Moreover, our proposed method also reduces the non-trivial robustness bias among different demographic groups, preventing the under-represented demographic groups from higher adversarial threats.
KW - Adversarial robustness
KW - Fair machine learning
KW - Human face recognition
UR - http://www.scopus.com/inward/record.url?scp=85147606173&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85147606173&partnerID=8YFLogxK
U2 - 10.1016/j.knosys.2023.110356
DO - 10.1016/j.knosys.2023.110356
M3 - Article
AN - SCOPUS:85147606173
SN - 0950-7051
VL - 264
JO - Knowledge-Based Systems
JF - Knowledge-Based Systems
M1 - 110356
ER -