当前位置:网站首页>iFair: Learning Individually Fair Data Representations for Algorithmic Decision Making
iFair: Learning Individually Fair Data Representations for Algorithmic Decision Making
2022-07-19 01:57:00 【Chubby Zhu】
subject : Learn individual fair data representation for algorithmic decision making
This paper introduces a method of mapping user record probability to low rank representation , This method coordinates the fairness of individuals and the utility of classifiers and rankings in downstream applications . Learning low rank data representation by defining a combined loss function , This work solves the basic trade-off between utility and fairness . The author applies their method to the classification of various real data sets and learns the ranking task , Proved the universality of their method . Their experiments show that , Compared with the best preliminary work of this setting , There are substantial improvements .( low-rank : The rank of the matrix is low ,eg:r=1)
Research background : In general, group equity is considered , Then the trade-off between personal fairness and accuracy has become a problem to be solved . here , A loss function is proposed to solve .
Research methods : Map an individual to an input space , This input space is independent of a specific protected subgroup . Define mapping functions , The sum of fairness loss and accuracy loss is expressed as the objective function . This research method is called iFair frame .
Personal fair representation and calculation : When there are two user records with all attributes xi and xj When the difference between the distance between and two records without sensitive attributes is less than or equal to the threshold , It means that there is personal fairness between these two user records .


The transformation mapping is expressed as : among ,U Represents the previously recorded probability distribution ,V Represents the prototype vector .


The loss function includes data loss and fairness loss :


Objective function :



experiment :5 Data sets
| Data sets | Protected properties | Output | Number of instances |
| ProPublica’s COMPAS | race | Commit another crime ? | |
| Census Income | gender | income >50K? | 48842 |
| German Credit | age | Credit value ? | 1000 |
| Airbnb | gender | Housing rating / Price ? | 27597 |
| gender | Job recommendation ranking ? |
The author randomly divided the data set into three parts . Use a part of the training model to learn model parameters , The second part is the verification set , Select the superparameter by performing a grid search ( See below for details ), The third part is the test set . The same data segmentation is used in this paper to compare all methods , Altogether 6 Methods .
| Method | Full Data | Masked Data | SVD | LFR | FA*IR( Sort ) | iFair |
| Data set used | the original dataset | the original dataset without protected attributes | full data/masked data | the learned representation | the representation learned by our model |
The evaluation index is personal fairness (yNN) And accuracy (ACC)

Experiments show that , In general iFair The effect of is better than other methods .
summary : Our method contains two important criteria . First , We look at fairness from the perspective of application independence , This allows us to incorporate it into various tasks , Including general classifiers and regression for learning to rank . secondly , We regard personal equity as an attribute of the dataset ( In a sense , Like Privacy ), This can be achieved by preprocessing the data into a converted representation . There is no need to access protected properties at this stage . If necessary , We can also post process the representation of learning , And implement the standards of group fairness , For example, statistical equivalence .
We apply this model to five real data sets , Prove that utility and personal fairness can be coordinated to a great extent . Apply classifier and regression model to iFair In the middle , The resulting algorithm decision is more consistent than the decision made on the original data . Our method is the first method to calculate the fair result separately in the learning sorting task . For classification tasks , Its performance is better than the existing advanced work .
边栏推荐
- The platform of digital collection NFT is good
- Boost thread pool
- nodejs-uuid
- 雾计算中的数据安全问题综述
- Handling Conditional Discrimination(可解释歧视和确切的歧视)
- windwos 下载安装OpenSSH
- [literature reading] tenet: a framework for modeling tensor dataflow based on relational centric notation
- Valgrind详细教程(1) Memcheck
- 深度伪造对国家安全的挑战及应对
- 【pycharm】Cannot find reference ‘XXX‘ in ‘__ init__. Py 'solution
猜你喜欢

动手学深度学习--线性神经网络篇

S32K148EVB 关于ENET Loopback实验

动手学深度学习---从全连接层到卷积层篇

VSCode中安装Go:tools failed to install.

L1,L2范数
![[literature reading] counting integer points in parametric polymers using barvinok's rational functions](/img/a2/3e1b248c7cd853ffea7a835111db65.png)
[literature reading] counting integer points in parametric polymers using barvinok's rational functions

微信公众号网页授权40029错误「建议收藏」

uniapp打包H5 空白页面 报错 Uncaught SyntaxError: Unexpected token ‘<‘

【文献阅读】Multi-state MRAM cells for hardware neuromorphic computing

记录一次海外图片加载不出来的排查
随机推荐
README.md添加目录
Laravel之文件上传
L1,L2范数
【MySQL】windows安装MySQL 5.7
iPhone 各大机型设备号
NFT differentiation trend has shown, how to capture value?
5G专网在智慧医疗中的应用
Learning Transferable Visual Models From Natural Language Supervision
S32K148EVB 关于ENET Loopback实验
mysql innodb 事务相关记录
基于移动互联网应用的儿童网络保护产业实践与完善
Valgrind detailed tutorial (1) MemCheck
Fairness in Semi-supervised Learning: Unlabeled Data Help to Reduce Discrimination
apt-get update报错:Hash 校验和不符
MXNet网络模型(五)条件GAN神经网络
[translation] transformers in computer vision
ROC 曲线讲解 (Receiver Operarating Curve)
CMTime简单介绍
Fair Attribute Classification through Latent Space De-biasing
Hands on deep learning -- linear neural network