Beyond Individual and Group Fairness

We present a new data-driven model of fairness that, unlike existing static definitions of individual or group fairness is guided by the unfairness complaints received by the system. Our model supports multiple fairness criteria and takes into account their potential incompatibilities. We consider both a stochastic and an adversarial setting of our model. In the stochastic setting, we show that our framework can be naturally cast as a Markov Decision Process with stochastic losses, for which we give efficient vanishing regret algorithmic solutions. In the adversarial setting, we design efficient algorithms with competitive ratio guarantees. We also report the results of experiments with our algorithms and the stochastic framework on artificial datasets, to demonstrate their effectiveness empirically.

[1]  Christopher T. Lowenkamp,et al.  False Positives, False Negatives, and False Analyses: A Rejoinder to "Machine Bias: There's Software Used across the Country to Predict Future Criminals. and It's Biased against Blacks" , 2016 .

[2]  Craig Boutilier,et al.  Optimizing Long-term Social Welfare in Recommender Systems: A Constrained Matching Approach , 2020, ICML.

[3]  Aditya Krishna Menon,et al.  Noise-tolerant fair classification , 2019, NeurIPS.

[4]  Aaron Roth,et al.  Fairness in Learning: Classic and Contextual Bandits , 2016, NIPS.

[5]  Avrim Blum,et al.  On preserving non-discrimination when combining expert advice , 2018, NeurIPS.

[6]  Silvio Lattanzi,et al.  Fair Clustering Through Fairlets , 2018, NIPS.

[7]  Esther Rolf,et al.  Delayed Impact of Fair Machine Learning , 2018, ICML.

[8]  John Langford,et al.  A Reductions Approach to Fair Classification , 2018, ICML.

[9]  Jon M. Kleinberg,et al.  On Fairness and Calibration , 2017, NIPS.

[10]  Kilian Q. Weinberger,et al.  On Calibration of Modern Neural Networks , 2017, ICML.

[11]  Aditya Krishna Menon,et al.  The cost of fairness in binary classification , 2018, FAT.

[12]  Julie A. Shah,et al.  Fairness in Multi-Agent Sequential Decision-Making , 2014, NIPS.

[13]  Aaron Roth,et al.  Meritocratic Fairness for Cross-Population Selection , 2017, ICML.

[14]  Ufuk Topcu,et al.  Fairness with Dynamics , 2019, ArXiv.

[15]  Maya R. Gupta,et al.  Training Well-Generalizing Classifiers for Fairness Metrics and Other Data-Dependent Constraints , 2018, ICML.

[16]  Cynthia Dwork,et al.  Individual Fairness in Pipelines , 2020, FORC.

[17]  Jon M. Kleinberg,et al.  Inherent Trade-Offs in the Fair Determination of Risk Scores , 2016, ITCS.

[18]  Aaron Roth,et al.  Fairness in Reinforcement Learning , 2016, ICML.

[19]  Jon M. Kleinberg,et al.  Selection Problems in the Presence of Implicit Bias , 2018, ITCS.

[20]  Ed H. Chi,et al.  Fairness in Recommendation Ranking through Pairwise Comparisons , 2019, KDD.

[21]  Harikrishna Narasimhan,et al.  Pairwise Fairness for Ranking and Regression , 2019, AAAI.

[22]  Christian Sohler,et al.  Fair Coresets and Streaming Algorithms for Fair k-Means Clustering , 2018, ArXiv.

[23]  Seth Neel,et al.  Preventing Fairness Gerrymandering: Auditing and Learning for Subgroup Fairness , 2017, ICML.

[24]  Michael I. Jordan,et al.  Robust Optimization for Fairness with Noisy Protected Groups , 2020, NeurIPS.

[25]  Philip S. Thomas,et al.  Importance Sampling for Fair Policy Selection , 2017, UAI.

[26]  Toniann Pitassi,et al.  Learning Fair Representations , 2013, ICML.

[27]  Aaron Roth,et al.  Average Individual Fairness: Algorithms, Generalization and Experiments , 2019, NeurIPS.

[28]  Matt J. Kusner,et al.  Counterfactual Fairness , 2017, NIPS.

[29]  Sampath Kannan,et al.  Downstream Effects of Affirmative Action , 2018, FAT.

[30]  Toniann Pitassi,et al.  Fairness through awareness , 2011, ITCS '12.

[31]  Nathan Srebro,et al.  From Fair Decision Making To Social Equality , 2018, FAT.

[32]  Nathan Srebro,et al.  Equality of Opportunity in Supervised Learning , 2016, NIPS.

[33]  Peter Auer,et al.  Near-optimal Regret Bounds for Reinforcement Learning , 2008, J. Mach. Learn. Res..

[34]  David Simchi-Levi,et al.  Phase Transitions in Bandits with Switching Constraints , 2019, NeurIPS.

[35]  Seth Neel,et al.  Meritocratic Fairness for Infinite and Contextual Bandits , 2018, AIES.

[36]  Nicolò Cesa-Bianchi,et al.  Online Learning with Switching Costs and Other Adaptive Adversaries , 2013, NIPS.

[37]  Toniann Pitassi,et al.  Predict Responsibly: Improving Fairness and Accuracy by Learning to Defer , 2017, NeurIPS.

[38]  Kush R. Varshney,et al.  Fair Transfer Learning with Missing Protected Attributes , 2019, AIES.

[39]  Adam Tauman Kalai,et al.  Decoupled Classifiers for Group-Fair and Efficient Machine Learning , 2017, FAT.

[40]  Percy Liang,et al.  Fairness Without Demographics in Repeated Loss Minimization , 2018, ICML.

[41]  Nisheeth K. Vishnoi,et al.  Ranking with Fairness Constraints , 2017, ICALP.

[42]  Peter Auer,et al.  Finite-time Analysis of the Multiarmed Bandit Problem , 2002, Machine Learning.

[43]  Krzysztof Onak,et al.  Scalable Fair Clustering , 2019, ICML.

[44]  Karthik Sridharan,et al.  Two-Player Games for Efficient Non-Convex Constrained Optimization , 2018, ALT.

[45]  Yang Liu,et al.  Calibrated Fairness in Bandits , 2017, ArXiv.

[46]  Christopher Jung,et al.  Online Learning with an Unknown Fairness Metric , 2018, NeurIPS.

[47]  Vijay Kamble,et al.  Individual Fairness in Hindsight , 2018, EC.