摘要
近几年深度神经网络正被广泛应用于现实决策系统,决策系统中的不公平现象会加剧社会不平等,造成社会危害.因此研究者们开始对深度学习系统的公平性展开大量研究,但大部分研究都从群体公平的角度切入,且这些缓解群体偏见的方法无法保证群体内部的公平.针对以上问题,定义两种个体公平率计算方法,分别为基于输出标签的个体公平率(IFRb),即相似样本对在模型预测中标签相同的概率和基于输出分布的个体公平率(IFRp),即相似样本对的预测分布差异在阈值范围内的概率,后者是更严格的个体公平.更进一步,提出一种提高模型个体公平性的算法IIFR,该算法通过余弦相似度计算样本之间的差异程度,利用相似临界值筛选出满足条件的相似训练样本对,最后在训练过程中将相似训练样本对的输出差异作为个体公平损失项添加到目标函数中,惩罚模型输出差异过大的相似训练样本对,以达到提高模型个体公平性的目的.实验结果表明, IIFR算法在个体公平的提升上优于最先进的个体公平提升方法.此外IIFR算法能够在提高模型个体公平性的同时,较好地维持模型的群体公平性.
In recent years,deep neural networks have been widely employed in real decision-making systems.Unfairness in decisionmaking systems will exacerbate social inequality and harm society.Therefore,researchers begin to carry out a lot of studies on the fairness of deep learning systems,where as most studies focus on group fairness and cannot guarantee fairness within the group.To this end,this study defines two individual fairness calculation methods.The first one is individual fairness rate IFRb based on labels of output,which is the probability of having the same predicted label for two similar samples.The second is individual fairness rate IFRp based on distributions of output,which is the probability of having similar predicted output distribution for two similar samples respectively,and the latter has stricter individual fairness.In addition,this study proposes an algorithm IIFR to improve the individual fairness of these models.The algorithm employs cosine similarity to measure the similarity between samples and then selects similar sample pairs via the similarity threshold decided by different applications.Finally,the output difference of the similar sample pairs is added to the objective function as an individual fairness loss item during the training,which penalizes the similar training samples with large differences in model output to improve the individual fairness of the model.The experimental results show that the proposed IIFR algorithm outperforms the state-of-the-art methods on individual fairness improvement,and can maintain group fairness of models while improving individual fairness.
作者
王昱颖
张敏
杨晶然
徐晟恺
陈仪香
WANG Yu-Ying;ZHANG Min;YANG Jing-Ran;XU Sheng-Kai;CHEN Yi-Xiang(Software Engineering Institute,East China Normal University,Shanghai 200062,China;Shanghai Key Laboratory of Trustworthy Computing(East China Normal University),Shanghai 200062,China)
出处
《软件学报》
EI
CSCD
北大核心
2023年第9期4037-4055,共19页
Journal of Software
基金
国家自然科学基金(61672012)
国家自然科学基金中以国际合作项目(62161146001)
科技部重点研发项目(2020AAA0107800)。
关键词
深度学习
模型偏见
个体公平
群体公平
deep learning
model bias
individual fairness
group fairness