Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3340531.3412705acmconferencesArticle/Chapter ViewAbstractPublication PagescikmConference Proceedingsconference-collections
research-article

LiFT: A Scalable Framework for Measuring Fairness in ML Applications

Published: 19 October 2020 Publication History

Abstract

Many internet applications are powered by machine learned models, which are usually trained on labeled datasets obtained through user feedback signals or human judgments. Since societal biases may be present in the generation of such datasets, it is possible for the trained models to be biased, thereby resulting in potential discrimination and harms for disadvantaged groups. Motivated by the need to understand and address algorithmic bias in web-scale ML systems and the limitations of existing fairness toolkits, we present the LinkedIn Fairness Toolkit (LiFT), a framework for scalable computation of fairness metrics as part of large ML systems. We highlight the key requirements in deployed settings, and present the design of our fairness measurement system. We discuss the challenges encountered in incorporating fairness tools in practice and the lessons learned during deployment at LinkedIn. Finally, we provide open problems based on practical experience.

Supplementary Material

MP4 File (3340531.3412705.mp4)
Many internet applications are powered by machine learned models, which are usually trained on labeled datasets obtained through either implicit / explicit user feedback signals or human judgments. Since societal biases may be present in the generation of such datasets, it is possible for the trained models to be biased, thereby resulting in potential discrimination and harms for disadvantaged groups. Motivated by the need for understanding and addressing algorithmic bias in web-scale ML systems and the limitations of existing fairness toolkits, we present the LinkedIn Fairness Toolkit (LiFT), a framework for scalable computation of fairness metrics as part of large ML systems. We highlight the key requirements in deployed settings, and present the design of our fairness measurement system. We discuss the challenges encountered in incorporating fairness tools in practice and the lessons learned during deployment at LinkedIn. Finally, we provide open problems based on practical experience.

References

[1]
M. Abadi et al. TensorFlow: A system for large-scale machine learning. In OSDI, 2016.
[2]
J. A. Adebayo. FairML: ToolBox for diagnosing bias in predictive modeling. PhD thesis, Massachusetts Institute of Technology, 2016.
[3]
A. Agarwal, A. Beygelzimer, M. Dudík, J. Langford, and H.Wallach. A reductions approach to fair classification. In ICML, 2018.
[4]
M. Armbrust et al. Spark SQL: Relational data processing in Spark. In SIGMOD, 2015.
[5]
N. Bantilan. Themis-ML: A fairness-aware machine learning interface for end-toend discrimination discovery and mitigation. Journal of Technology in Human Services, 36(1), 2018.
[6]
S. Barocas and M. Hardt. Fairness in machine learning. In NeurIPS Tutorial, 2017.
[7]
R. K. Bellamy et al. AI Fairness 360: An extensible toolkit for detecting, understanding, and mitigating unwanted algorithmic bias. arXiv preprint arXiv:1810.01943, 2018.
[8]
A. Beutel, J. Chen, T. Doshi, H. Qian, A.Woodruff, C. Luu, P. Kreitmann, J. Bischof, and E. H. Chi. Putting fairness principles into practice: Challenges, metrics, and improvements. In AIES, 2019.
[9]
A. Chouldechova. Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. Big data, 5(2), 2017.
[10]
K. Crawford. The trouble with bias. In NeurIPS Invited Talk, 2017.
[11]
C. DiCiccio, S. Vasudevan, K. Basu, K. Kenthapadi, and D. Agarwal. Evaluating fairness using permutation tests. In KDD, 2020.
[12]
S. A. Friedler, C. Scheidegger, S. Venkatasubramanian, S. Choudhary, E. P. Hamilton, and D. Roth. A comparative study of fairness-enhancing interventions in machine learning. In FAT*, 2019.
[13]
S. Galhotra, Y. Brun, and A. Meliou. Fairness testing: Testing software for discrimination. In ESEC/FSE, 2017.
[14]
S. C. Geyik, S. Ambler, and K. Kenthapadi. Fairness-aware ranking in search & recommendation systems with application to LinkedIn talent search. In KDD, 2019.
[15]
P. Good. Permutation Tests: A Practical Guide to Resampling Methods for Testing Hypotheses. Springer series in statistics. Springer, 2000.
[16]
B. Green and Y. Chen. Disparate interactions: An algorithm-in-the-loop analysis of fairness in risk assessments. In FAT*, 2019.
[17]
N. Grgic-Hlaca, E. M. Redmiles, K. P. Gummadi, and A. Weller. Human perceptions of fairness in algorithmic decision making: A case study of criminal risk prediction. In WWW, 2018.
[18]
M. Hardt, E. Price, and N. Srebro. Equality of opportunity in supervised learning. In NIPS, 2016.
[19]
K. Holstein, J. Wortman Vaughan, H. Daumé III, M. Dudik, and H. Wallach. Improving fairness in machine learning systems: What do industry practitioners need? In CHI, 2019.
[20]
F. Kamiran and T. Calders. Data preprocessing techniques for classification without discrimination. Knowledge and Information Systems, 33(1):1--33, 2012.
[21]
T. Kamishima, S. Akaho, H. Asoh, and J. Sakuma. Fairness-aware classifier with prejudice remover regularizer. In ECML PKDD, 2012.
[22]
R. Kohavi. Scaling up the accuracy of Naive-Bayes classifiers: A decision-tree hybrid. In KDD, 1996.
[23]
M. Mitchell, S.Wu, A. Zaldivar, P. Barnes, L. Vasserman, B. Hutchinson, E. Spitzer, I. D. Raji, and T. Gebru. Model cards for model reporting. In FAT*, 2019.
[24]
M. Ojala and G. C. Garriga. Permutation tests for studying classifier performance. Journal of Machine Learning Research, 11, 2010.
[25]
F. Pedregosa et al. Scikit-learn: Machine learning in Python. Journal of Machine Learning Research, 12, 2011.
[26]
pymetrics. audit-AI, 2018. https://github.com/pymetrics/audit-ai.
[27]
P. Saleiro, B. Kuester, A. Stevens, A. Anisfeld, L. Hinkson, J. London, and R. Ghani. Aequitas: A bias and fairness audit toolkit. arXiv preprint arXiv:1811.05577, 2018.
[28]
T. Speicher, H. Heidari, N. Grgic-Hlaca, K. P. Gummadi, A. Singla, A. Weller, and M. B. Zafar. A unified approach to quantifying algorithmic unfairness: Measuring individual & group unfairness via inequality indices. In KDD, 2018.
[29]
M. Srivastava, H. Heidari, and A. Krause. Mathematical notions vs. human perception of fairness: A descriptive approach to fairness for machine learning. In KDD, 2019.
[30]
A. Swami, S. Vasudevan, and J. Huyn. Data sentinel: A declarative production scale data validation platform. In ICDE, 2020.
[31]
F. Tramer, V. Atlidakis, R. Geambasu, D. Hsu, J.-P. Hubaux, M. Humbert, A. Juels, and H. Lin. FairTest: Discovering unwarranted associations in data-driven applications. In Euro S&P, 2017.
[32]
M. Zaharia et al. Apache spark: a unified engine for big data processing. Communications of the ACM, 59(11), 2016.
[33]
M. Zehlike, C. Castillo, F. Bonchi, S. Hajian, and M. Megahed. Fairness Measures: Datasets and software for detecting algorithmic discrimination, 2017. http://fairness-measures.org/.

Cited By

View all
  • (2024)Fairness Testing: A Comprehensive Survey and Analysis of TrendsACM Transactions on Software Engineering and Methodology10.1145/365215533:5(1-59)Online publication date: 4-Jun-2024
  • (2024)PreFAIR: Combining Partial Preferences for Fair Consensus Decision-makingProceedings of the 2024 ACM Conference on Fairness, Accountability, and Transparency10.1145/3630106.3658961(1133-1149)Online publication date: 3-Jun-2024
  • (2024)A Scoping Study of Evaluation Practices for Responsible AI Tools: Steps Towards Effectiveness EvaluationsProceedings of the 2024 CHI Conference on Human Factors in Computing Systems10.1145/3613904.3642398(1-24)Online publication date: 11-May-2024
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
CIKM '20: Proceedings of the 29th ACM International Conference on Information & Knowledge Management
October 2020
3619 pages
ISBN:9781450368599
DOI:10.1145/3340531
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 19 October 2020

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. distributed computation
  2. fairness-aware machine learning
  3. linkedin fairness toolkit
  4. scalable framework

Qualifiers

  • Research-article

Conference

CIKM '20
Sponsor:

Acceptance Rates

Overall Acceptance Rate 1,861 of 8,427 submissions, 22%

Upcoming Conference

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)48
  • Downloads (Last 6 weeks)3
Reflects downloads up to 15 Oct 2024

Other Metrics

Citations

Cited By

View all
  • (2024)Fairness Testing: A Comprehensive Survey and Analysis of TrendsACM Transactions on Software Engineering and Methodology10.1145/365215533:5(1-59)Online publication date: 4-Jun-2024
  • (2024)PreFAIR: Combining Partial Preferences for Fair Consensus Decision-makingProceedings of the 2024 ACM Conference on Fairness, Accountability, and Transparency10.1145/3630106.3658961(1133-1149)Online publication date: 3-Jun-2024
  • (2024)A Scoping Study of Evaluation Practices for Responsible AI Tools: Steps Towards Effectiveness EvaluationsProceedings of the 2024 CHI Conference on Human Factors in Computing Systems10.1145/3613904.3642398(1-24)Online publication date: 11-May-2024
  • (2024) Implementing equitable and intersectionality‐aware ML in education: A practical guide British Journal of Educational Technology10.1111/bjet.1348455:5(2003-2038)Online publication date: 23-May-2024
  • (2024)Linking convolutional kernel size to generalization bias in face analysis CNNs2024 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)10.1109/WACV57701.2024.00464(4693-4703)Online publication date: 3-Jan-2024
  • (2024)Individual Fairness with Group Awareness Under UncertaintyMachine Learning and Knowledge Discovery in Databases. Research Track10.1007/978-3-031-70362-1_6(89-106)Online publication date: 22-Aug-2024
  • (2024)AI fairness in practice: Paradigm, challenges, and prospectsAI Magazine10.1002/aaai.12189Online publication date: 22-Sep-2024
  • (2023)Algorithmic FairnessAnnual Review of Financial Economics10.1146/annurev-financial-110921-12593015:1(565-593)Online publication date: 1-Nov-2023
  • (2023)“☑ Fairness Toolkits, A Checkbox Culture?” On the Factors that Fragment Developer Practices in Handling Algorithmic HarmsProceedings of the 2023 AAAI/ACM Conference on AI, Ethics, and Society10.1145/3600211.3604674(482-495)Online publication date: 8-Aug-2023
  • (2023)Disentangling and Operationalizing AI Fairness at LinkedInProceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency10.1145/3593013.3594075(1213-1228)Online publication date: 12-Jun-2023
  • Show More Cited By

View Options

Get Access

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media