论文标题
公平概念和相关紧张局势的调查
Survey on Fairness Notions and Related Tensions
论文作者
论文摘要
自动化决策系统越来越多地用于在诸如雇用和贷款授予等问题的问题上做出结果决策,以期通过客观机器学习(ML)算法代替主观人类决策。但是,基于ML的决策系统容易出现偏见,这导致了不公平的决策。文献中已经定义了几种公平概念,以捕获这种道德和社会概念的不同微妙之处(例如统计奇偶校验,均等机会等)。当学习模型时,要满足的公平要求在不同的公平概念和其他期望的属性(例如隐私和分类精度)之间产生了几种类型的紧张局势。本文调查了常用的公平概念,并以隐私和准确性讨论了它们之间的紧张局势。审查了解决公平准确性权衡的不同方法(分类为四种方法,即预处理,进行内部处理,后处理和混合动力)。该调查通过在公平基准数据集上进行的实验分析进行了整合,以说明在现实情况下的公平度量与准确性之间的关系。
Automated decision systems are increasingly used to take consequential decisions in problems such as job hiring and loan granting with the hope of replacing subjective human decisions with objective machine learning (ML) algorithms. However, ML-based decision systems are prone to bias, which results in yet unfair decisions. Several notions of fairness have been defined in the literature to capture the different subtleties of this ethical and social concept (e.g., statistical parity, equal opportunity, etc.). Fairness requirements to be satisfied while learning models created several types of tensions among the different notions of fairness and other desirable properties such as privacy and classification accuracy. This paper surveys the commonly used fairness notions and discusses the tensions among them with privacy and accuracy. Different methods to address the fairness-accuracy trade-off (classified into four approaches, namely, pre-processing, in-processing, post-processing, and hybrid) are reviewed. The survey is consolidated with experimental analysis carried out on fairness benchmark datasets to illustrate the relationship between fairness measures and accuracy in real-world scenarios.