Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3506860.3506902acmotherconferencesArticle/Chapter ViewAbstractPublication PageslakConference Proceedingsconference-collections
research-article
Open access

FairEd: A Systematic Fairness Analysis Approach Applied in a Higher Educational Context

Published: 21 March 2022 Publication History

Abstract

Higher education institutions increasingly rely on machine learning models. However, a growing body of evidence shows that these algorithms may not serve underprivileged communities well and at times discriminate against them. This is all the more concerning in education as negative outcomes have long-term implications. We propose a systematic process for framing, detecting, documenting, and reporting unfairness risks. The systematic approach’s outcomes are merged into a framework named FairEd, which would help decision-makers to understand unfairness risks along the environmental and analytical fairness dimension. The tool allows to decide (i) whether the dataset contains risks of unfairness; (ii) how the models could perform along many fairness dimensions; (iii) whether potentially unfair outcomes can be mitigated without degrading performance. The systematic approach is applied to a Chilean University case study, where a predicting student dropout model is aimed to build. First, we capture the nuances of the Chilean context where unfairness emerges along income lines and demographic groups. Second, we highlight the benefit of reporting unfairness risks along a diverse set of metrics to shed light on potential discrimination. Third, we find that measuring the cost of fairness is an important quantity to report on when doing the model selection.

References

[1]
Philip Adler, Casey Falk, Sorelle A Friedler, Tionney Nix, Gabriel Rybeck, Carlos Scheidegger, Brandon Smith, and Suresh Venkatasubramanian. 2018. Auditing black-box models for indirect influence. Knowledge and Information Systems 54, 1 (2018), 95–122.
[2]
Alekh Agarwal, Alina Beygelzimer, Miroslav Dudík, John Langford, and Hanna Wallach. 2018. A reductions approach to fair classification. arXiv preprint arXiv:1803.02453(2018).
[3]
Lovenoor Aulck, Dev Nambi, Nishant Velagapudi, Joshua Blumenstock, and Jevin West. 2019. Mining University Registrar Records to Predict First-Year Undergraduate Attrition.International Educational Data Mining Society (2019).
[4]
Ryan S. Baker and Aaron Hawn. 2021. lgorithmic Bias in Education. EdArXiv (2021).
[5]
John P Bean. 1980. Dropouts and turnover: The synthesis and test of a causal model of student attrition. Research in higher education 12, 2 (1980), 155–187.
[6]
John P Bean and Barbara S Metzner. 1985. A conceptual model of nontraditional undergraduate student attrition. Review of educational Research 55, 4 (1985), 485–540.
[7]
Rachel K. E. Bellamy, Kuntal Dey, Michael Hind, Samuel C. Hoffman, Stephanie Houde, Kalapriya Kannan, Pranay Lohia, Jacquelyn Martino, Sameep Mehta, Aleksandra Mojsilovic, Seema Nagar, Karthikeyan Natesan Ramamurthy, John Richards, Diptikalyan Saha, Prasanna Sattigeri, Moninder Singh, Kush R. Varshney, and Yunfeng Zhang. 2018. AI Fairness 360: An Extensible Toolkit for Detecting, Understanding, and Mitigating Unwanted Algorithmic Bias. https://arxiv.org/abs/1810.01943
[8]
Emily M Bender and Batya Friedman. 2018. Data statements for natural language processing: Toward mitigating system bias and enabling better science. Transactions of the Association for Computational Linguistics 6 (2018), 587–604.
[9]
Trudy H Bers and Kerry E Smith. 1991. Persistence of community college students: The influence of student intent and academic and social integration. Research in higher Education 32, 5 (1991), 539–556.
[10]
Sarah Bird, Miro Dudík, Richard Edgar, Brandon Horn, Roman Lutz, Vanessa Milan, Mehrnoosh Sameki, Hanna Wallach, and Kathleen Walker. 2020. Fairlearn: A toolkit for assessing and improving fairness in AI. Microsoft, Tech. Rep. MSR-TR-2020-32(2020).
[11]
John M Braxton, Ana V Shaw Sullivan, and Robert M Johnson. 1997. Appraising Tinto’s theory of college student departure. HIGHER EDUCATION-NEW YORK-AGATHON PRESS INCORPORATED- 12 (1997), 107–164.
[12]
Joy Buolamwini and Timnit Gebru. 2018. Gender Shades: Intersectional Accuracy Disparities in Commercial Gender Classification. In Proceedings of the 1st Conference on Fairness, Accountability and Transparency(Proceedings of Machine Learning Research, Vol. 81), Sorelle A. Friedler and Christo Wilson (Eds.). PMLR, New York, NY, USA, 77–91. http://proceedings.mlr.press/v81/buolamwini18a.html
[13]
Alberto F Cabrera, Amaury Nora, and Maria B Castaneda. 1993. College persistence: Structural equations modeling test of an integrated model of student retention. The journal of higher education 64, 2 (1993), 123–139.
[14]
Flavio Calmon, Dennis Wei, Bhanukiran Vinzamuri, Karthikeyan Natesan Ramamurthy, and Kush R Varshney. 2017. Optimized Pre-Processing for Discrimination Prevention. In Advances in Neural Information Processing Systems 30, I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett (Eds.). Curran Associates, Inc., 3992–4001.
[15]
Hongyan Chang and Reza Shokri. 2020. On the Privacy Risks of Algorithmic Fairness. arXiv preprint arXiv:2011.03731(2020).
[16]
Alexandra Chouldechova and Aaron Roth. 2018. The frontiers of fairness in machine learning. arXiv preprint arXiv:1810.08810(2018).
[17]
Kristof Coussement, Minh Phan, Arno De Caigny, Dries F. Benoit, and Annelies Raes. 2020. Predicting student dropout in subscription-based online learning environments: The beneficial impact of the logit leaf model. Decision Support Systems 135 (2020), 113325.
[18]
Deana B Davalos, Ernest L Chavez, and Robert J Guardiola. 1999. The effects of extracurricular activity, ethnic identification, and perception of school on student dropout rates. Hispanic Journal of behavioral sciences 21, 1 (1999), 61–77.
[19]
Consejo Nacional de Educacion. [n. d.]. Retencion Primer Año. https://www.cned.cl/indices/retencion-primer-ano. Accessed: 2020-11-30.
[20]
Dursun Delen. 2011. Predicting student attrition with data mining methods. Journal of College Student Retention: Research, Theory & Practice 13, 1(2011), 17–35.
[21]
Cynthia Dwork, Moritz Hardt, Toniann Pitassi, Omer Reingold, and Richard Zemel. 2012. Fairness through awareness. In Proceedings of the 3rd innovations in theoretical computer science conference. 214–226.
[22]
Harrison Edwards and Amos Storkey. 2015. Censoring representations with an adversary. arXiv preprint arXiv:1511.05897(2015).
[23]
Gregory Elacqua. 2012. The impact of school choice and public policy on segregation: Evidence from Chile. International Journal of Educational Development 32, 3(2012), 444–453.
[24]
Michael Feldman, Sorelle A Friedler, John Moeller, Carlos Scheidegger, and Suresh Venkatasubramanian. 2015. Certifying and removing disparate impact. In proceedings of the 21th ACM SIGKDD international conference on knowledge discovery and data mining. 259–268.
[25]
Antonio Jesús Fernández-García, Roberto Rodríguez-Echeverría, Juan Carlos Preciado, José María Conejero Manzano, and Fernando Sánchez-Figueroa. 2020. Creating a Recommender System to Support Higher Education Students in the Subject Enrollment Decision. IEEE Access 8(2020), 189069–189088.
[26]
Josep Figueroa-Cañas and Teresa Sancho-Vinuesa. 2020. Early Prediction of Dropout and Final Exam Performance in an Online Statistics Course. IEEE Revista Iberoamericana de Tecnologias del Aprendizaje 15, 2(2020), 86–94.
[27]
National Center for Education Statistics. 2020. Undergraduate retention and graduation rates. (2020).
[28]
Sorelle A Friedler, Carlos Scheidegger, and Suresh Venkatasubramanian. 2021. The (im) possibility of fairness: Different value systems require different mechanisms for fair decision making. Commun. ACM 64, 4 (2021), 136–143.
[29]
Josh Gardner, Christopher Brooks, and Ryan Baker. 2019. Evaluating the fairness of predictive student models through slicing analysis. In Proceedings of the 9th International Conference on Learning Analytics & Knowledge. 225–234.
[30]
Timnit Gebru, Jamie Morgenstern, Briana Vecchione, Jennifer Wortman Vaughan, Hanna Wallach, Hal Daumé III, and Kate Crawford. 2018. Datasheets for datasets. arXiv preprint arXiv:1803.09010(2018).
[31]
Xavier Gitiaux and Huzefa Rangwala. 2019. mdfa: Multi-Differential Fairness Auditor for Black Box Classifiers. In IJCAI. 5871–5879.
[32]
Paula Gordaliza, Eustasio Del Barrio, Gamboa Fabrice, and Jean-Michel Loubes. 2019. Obtaining fairness using optimal transport theory. In International Conference on Machine Learning. 2357–2365.
[33]
Nina Grgic-Hlaca, Muhammad Bilal Zafar, Krishna P Gummadi, and Adrian Weller. 2016. The case for process fairness in learning: Feature selection for fair decision making. In NIPS Symposium on Machine Learning and the Law, Vol. 1. 2.
[34]
Lani Guinier. 2015. The tyranny of the meritocracy: Democratizing higher education in America. Beacon Press.
[35]
Moritz Hardt, Eric Price, and Nathan Srebro. 2016. Equality of opportunity in supervised learning. arXiv preprint arXiv:1610.02413(2016).
[36]
Irina Higgins, Loic Matthey, Arka Pal, Christopher Burgess, Xavier Glorot, Matthew Botvinick, Shakir Mohamed, and Alexander Lerchner. 2016. beta-vae: Learning basic visual concepts with a constrained variational framework. (2016).
[37]
Isabel Hilliger, Margarita Ortiz-Rojas, Paola Pesántez-Cabrera, Eliana Scheihing, Yi-Shan Tsai, Pedro J Muñoz-Merino, Tom Broos, Alexander Whitelock-Wainwright, and Mar Pérez-Sanagustín. 2020. Identifying needs for learning analytics adoption in Latin American universities: A mixed-methods approach. The Internet and Higher Education 45 (2020), 100726.
[38]
Qian Hu and Huzefa Rangwala. 2020. Metric-Free Individual Fairness with Cooperative Contextual Bandits. arXiv preprint arXiv:2011.06738(2020).
[39]
Qian Hu and Huzefa Rangwala. 2020. Towards Fair Educational Data Mining: A Case Study on Detecting At-risk Students. In Proceedings of The 13th International Conference on Educational Data Mining (EDM 2020). 431–437.
[40]
Huade Huo, Jiashan Cui, Sarah Hein, Zoe Padgett, Mark Ossolinski, Ruth Raim, and Jijun Zhang. 2020. Predicting Dropout for Nontraditional Undergraduate Students: A Machine Learning Approach. Journal of College Student Retention: Research, Theory & Practice (2020), 1521025120963821.
[41]
Stephen Hutt, Margo Gardner, Angela L Duckworth, and Sidney K D’Mello. 2019. Evaluating Fairness and Generalizability in Models Predicting On-Time Graduation from College Applications.International Educational Data Mining Society (2019).
[42]
Sajad Khodadadian, AmirEmad Ghassami, and Negar Kiyavash. 2021. Impact of Data Processing on Fairness in Supervised Learning. arXiv preprint arXiv:2102.01867(2021).
[43]
René F Kizilcec and Hansol Lee. 2020. Algorithmic Fairness in Education. arXiv preprint arXiv:2007.05443(2020).
[44]
Jon Kleinberg, Sendhil Mullainathan, and Manish Raghavan. 2016. Inherent trade-offs in the fair determination of risk scores. arXiv preprint arXiv:1609.05807(2016).
[45]
Matt J Kusner, Joshua Loftus, Chris Russell, and Ricardo Silva. 2017. Counterfactual fairness. In Advances in neural information processing systems. 4066–4076.
[46]
David Kynaston and Francis Green. 2019. Engines of privilege: Britain’s private school problem. Bloomsbury Publishing Plc.
[47]
Dean R Lillard and Philip P DeCicca. 2001. Higher standards, more dropouts? Evidence within and across time. Economics of Education Review 20, 5 (2001), 459–473.
[48]
Christos Louizos, Kevin Swersky, Yujia Li, Max Welling, and Richard Zemel. 2015. The variational fair autoencoder. arXiv preprint arXiv:1511.00830(2015).
[49]
David Madras, Elliot Creager, Toniann Pitassi, and Richard Zemel. 2018. Learning adversarially fair and transferable representations. In International Conference on Machine Learning. PMLR, 3384–3393.
[50]
Charles T Marx, Richard Lanas Phillips, Sorelle A Friedler, Carlos Scheidegger, and Suresh Venkatasubramanian. 2019. Disentangling Influence: Using disentangled representations to audit model predictions. arXiv preprint arXiv:1906.08652(2019).
[51]
Margaret Mitchell, Simone Wu, Andrew Zaldivar, Parker Barnes, Lucy Vasserman, Ben Hutchinson, Elena Spitzer, Inioluwa Deborah Raji, and Timnit Gebru. 2019. Model cards for model reporting. In Proceedings of the conference on fairness, accountability, and transparency. 220–229.
[52]
Paul A Murtaugh, Leslie D Burns, and Jill Schuster. 1999. Predicting the retention of university students. Research in higher education 40, 3 (1999), 355–371.
[53]
Senthil Kumar Narayanasamy and Atilla Elçi. 2020. An effective prediction model for online course dropout rate. International Journal of Distance Education Technologies (IJDET) 18, 4(2020), 94–110.
[54]
Zachary A Pardos, Zihao Fan, and Weijie Jiang. 2019. Connectionist recommendation in the wild: on the utility and scrutability of neural networks for personalized course guidance. User modeling and user-adapted interaction 29, 2 (2019), 487–525.
[55]
Angie Parker. 1999. A study of variables that predict dropout from distance education. International journal of educational technology 1, 2 (1999), 1–10.
[56]
Fabian Pedregosa, Gaël Varoquaux, Alexandre Gramfort, Vincent Michel, Bertrand Thirion, Olivier Grisel, Mathieu Blondel, Peter Prettenhofer, Ron Weiss, Vincent Dubourg, 2011. Scikit-learn: Machine learning in Python. the Journal of machine Learning research 12 (2011), 2825–2830.
[57]
Stephen Pfohl, Ben Marafino, Adrien Coulet, Fatima Rodriguez, Latha Palaniappan, and Nigam H Shah. 2019. Creating Fair Models of Atherosclerotic Cardiovascular Disease Risk. In Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society. ACM, 271–278.
[58]
ProPublica. 2016. How We Analyzed the COMPAS Recidivism Algorithm. ProPublica (2016).
[59]
Nor Samsiah Sani, Ahmad Fikri Mohamed Nafuri, Zulaiha Ali Othman, Mohd Zakree Ahmad Nazri, and Khairul Nadiyah Mohamad. 2020. Drop-Out Prediction in Higher Education Among B40 Students. International Journal of Advanced Computer Science and Applications (2020).
[60]
William G Spady. 1970. Dropouts from higher education: An interdisciplinary review and synthesis. Interchange 1, 1 (1970), 64–85.
[61]
William G Spady. 1971. Dropouts from higher education: Toward an empirical model. Interchange 2, 3 (1971), 38–62.
[62]
Nate Sutter and Sharon Paulson. 2017. Predicting college students’ intention to graduate: a test of the theory of planned behavior. College Student Journal 50, 3 (2017), 409–421.
[63]
Mack Sweeney, Huzefa Rangwala, Jaime Lester, and Aditya Johri. 2016. Next-term student performance prediction: A recommender systems approach. arXiv preprint arXiv:1604.01840(2016).
[64]
Vincent Tinto. 1975. Dropout from higher education: A theoretical synthesis of recent research. Review of educational research 45, 1 (1975), 89–125.
[65]
Vincent Tinto. 2007. Taking student retention seriously. Syracuse University Syracuse, NY.
[66]
Vincent Tinto and John Cullen. 1973. Dropout in Higher Education: A Review and Theoretical Synthesis of Recent Research.(1973).
[67]
Jonathan Vásquez and Jaime Miranda. 2019. Student Desertion: What Is and How Can It Be Detected on Time?In Data Science and Digital Business. Springer, 263–283.
[68]
Muhammad Bilal Zafar, Isabel Valera, Manuel Rodriguez, Krishna Gummadi, and Adrian Weller. 2017. From parity to preference-based notions of fairness in classification. In Advances in Neural Information Processing Systems. 229–239.
[69]
Rich Zemel, Yu Wu, Kevin Swersky, Toni Pitassi, and Cynthia Dwork. 2013. Learning fair representations. In International Conference on Machine Learning. 325–333.
[70]
Seth D Zimmerman. 2019. Elite colleges and upward mobility to top jobs and top incomes. American Economic Review 109, 1 (2019), 1–47.

Cited By

View all
  • (2024)Educational Data Mining for Dropout Prediction: Trends, Opportunities, and ChallengesRevista Brasileira de Informática na Educação10.5753/rbie.2024.355932(220-256)Online publication date: 20-May-2024
  • (2024)Contexts Matter but How? Course-Level Correlates of Performance and Fairness Shift in Predictive Model TransferProceedings of the 14th Learning Analytics and Knowledge Conference10.1145/3636555.3636936(713-724)Online publication date: 18-Mar-2024
  • (2024)Investigating Algorithmic Bias on Bayesian Knowledge Tracing and Carelessness DetectorsProceedings of the 14th Learning Analytics and Knowledge Conference10.1145/3636555.3636890(349-359)Online publication date: 18-Mar-2024
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Other conferences
LAK22: LAK22: 12th International Learning Analytics and Knowledge Conference
March 2022
582 pages
ISBN:9781450395731
DOI:10.1145/3506860
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 21 March 2022

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. algorithmic fairness
  2. educational data mining
  3. student dropout

Qualifiers

  • Research-article
  • Research
  • Refereed limited

Conference

LAK22

Acceptance Rates

Overall Acceptance Rate 236 of 782 submissions, 30%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)354
  • Downloads (Last 6 weeks)55
Reflects downloads up to 10 Nov 2024

Other Metrics

Citations

Cited By

View all
  • (2024)Educational Data Mining for Dropout Prediction: Trends, Opportunities, and ChallengesRevista Brasileira de Informática na Educação10.5753/rbie.2024.355932(220-256)Online publication date: 20-May-2024
  • (2024)Contexts Matter but How? Course-Level Correlates of Performance and Fairness Shift in Predictive Model TransferProceedings of the 14th Learning Analytics and Knowledge Conference10.1145/3636555.3636936(713-724)Online publication date: 18-Mar-2024
  • (2024)Investigating Algorithmic Bias on Bayesian Knowledge Tracing and Carelessness DetectorsProceedings of the 14th Learning Analytics and Knowledge Conference10.1145/3636555.3636890(349-359)Online publication date: 18-Mar-2024
  • (2024)What Fairness Metrics Can Really Tell You: A Case Study in the Educational DomainProceedings of the 14th Learning Analytics and Knowledge Conference10.1145/3636555.3636873(792-799)Online publication date: 18-Mar-2024
  • (2024) Implementing equitable and intersectionality‐aware ML in education: A practical guide British Journal of Educational Technology10.1111/bjet.1348455:5(2003-2038)Online publication date: 23-May-2024
  • (2023)The landscape of data and AI documentation approaches in the European policy contextEthics and Information Technology10.1007/s10676-023-09725-725:4Online publication date: 28-Oct-2023
  • (2023)A Review of Clustering Models in Educational Data Science Toward Fairness-Aware LearningEducational Data Science: Essentials, Approaches, and Tendencies10.1007/978-981-99-0026-8_2(43-94)Online publication date: 30-Apr-2023

View Options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format.

HTML Format

Get Access

Login options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media