Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3303772.3303791acmotherconferencesArticle/Chapter ViewAbstractPublication PageslakConference Proceedingsconference-collections
research-article

Evaluating the Fairness of Predictive Student Models Through Slicing Analysis

Published: 04 March 2019 Publication History

Abstract

Predictive modeling has been a core area of learning analytics research over the past decade, with such models currently deployed in a variety of educational contexts from MOOCs to K-12. However, analyses of the differential effectiveness of these models across demographic, identity, or other groups has been scarce. In this paper, we present a method for evaluating unfairness in predictive student models. We define this in terms of differential accuracy between subgroups, and measure it using a new metric we term the Absolute Between-ROC Area (ABROCA). We demonstrate the proposed method through a gender-based "slicing analysis" using five different models replicated from other works and a dataset of 44 unique MOOCs and over four million learners. Our results demonstrate (1) significant differences in model fairness according to (a) statistical algorithm and (b) feature set used; (2) that the gender imbalance ratio, curricular area, and specific course used for a model all display significant association with the value of the ABROCA statistic; and (3) that there is not evidence of a strict tradeoff between performance and fairness. This work provides a framework for quantifying and understanding how predictive models might inadvertently privilege, or disparately impact, different student subgroups. Furthermore, our results suggest that learning analytics researchers and practitioners can use slicing analysis to improve model fairness without necessarily sacrificing performance.1

References

[1]
JML Andres, RS Baker, G Siemens, D Gašević, and S Crossley. 2018. Studying MOOC Completion at Scale Using the MOOC Replication Framework. In Proc. LAK. ACM, New York, 71--78.
[2]
V Bakthavachalam and A Hickey. 2018. Using Technology to Increase Equality of Opportunity. https://digital.hbs.edu/data-and-analysis/using-technology-to-increase-equality-of-opportunity/. (2018). Accessed: 2018-9-17.
[3]
Solon Barocas and Andrew D Selbst. 2016. Big data's disparate impact. Calif. Law Rev. 104 (2016), 671.
[4]
C G Brinton and M Chiang. 2015. MOOC performance prediction via clickstream data and social learning networks. In IEEE INFOCOM. IEEE, New York, 2299--2307.
[5]
Christopher Brooks, Joshua Gardner, and Kaifeng Chen. 2018. How Gender Cues in Educational Video Impact Participation and Retention. In Proc. ICLS, J Kay and R Luckin (Eds.). ISLS, New York, 1835--1842.
[6]
Isaac Chuang and Andrew Dean Ho. 2016. HarvardX and MITx: Four Years of Open Online Courses -- Fall 2012-Summer 2016. Technical Report. Harvard/MIT.
[7]
Kate Crawford and Ryan Calo. 2016. There is a blind spot in AI research. Nature News 538, 7625 (Oct. 2016), 311.
[8]
S Crossley, DS McNamara, R Baker, Y Wang, L Paquette, T Barnes, and Y Bergner. 2015. Language to Completion: Success in an Educational Data Mining Massive Open Online Class. In Proc. EDM. 388--391.
[9]
Cynthia Dwork, Moritz Hardt, Toniann Pitassi, Omer Reingold, and Richard Zemel. 2012. Fairness Through Awareness. In Proc. ITCS. ACM, 214--226.
[10]
C Dwork, NImmorlica, AT Kalai, and MDM Leiserson. 2018. Decoupled Classifiers for Group-Fair and Efficient Machine Learning. In Proc. FAT*, SA Friedler and C Wilson (Eds.), Vol. 81. PMLR, New York, NY, USA, 119--133.
[11]
M Fei and D Y Yeung. 2015. Temporal Models for Predicting Student Dropout in Massive Open Online Courses. In ICDMW). IEEE, New York, 256--263.
[12]
M Feldman, SA Friedler, J Moeller, C Scheidegger, and S Venkatasubramanian. 2015. Certifying and Removing Disparate Impact. In Proc. KDD. ACM, 259--268.
[13]
SA Friedler, C Scheidegger, S Venkatasubramanian, S Choudhary, EP Hamilton, and D Roth. 2018. A comparative study of fairness-enhancing interventions in machine learning. (2018). arXiv:stat.ML/1802.04422
[14]
J Gardner and C Brooks. 2018. Evaluating Predictive Models of Student Success: Closing the Methodological Gap. Journal of Learning Analytics 5, 2 (2018), 105--125.
[15]
Josh Gardner and Christopher Brooks. 2018. Student Success Prediction in MOOCs. User Modeling and User-Adapted Interaction 28, 2 (2018), 127--203.
[16]
J Gardner, C Brooks, JML Andres, and R Baker. 2018. Replicating MOOC Predictive Models at Scale. In Proc. Learning@Scale. ACM, New York, 1--10.
[17]
Josh Gardner, Christopher Brooks, Juan Miguel L Andres, and Ryan Baker. 2018. MORF: A Framework for Predictive Modeling and Replication At Scale With Privacy-Restricted MOOC Data. (2018). arXiv:1801.05236
[18]
J A Hanley and B J McNeil. 1982. The meaning and use of the area under a receiver operating characteristic (ROC) curve. Radiology 143, 1 (1982), 29--36.
[19]
Moritz Hardt, Eric Price, Nati Srebro, and Others. 2016. Equality of opportunity in supervised learning. In NIPS. 3315--3323.
[20]
M Hind, S Mehta, A Mojsilovic, R Nair, K Natesan Ramamurthy, A Olteanu, and KR Varshney. 2018. Increasing Trust in AI Services through Supplier's Declarations of Conformity. (Aug. 2018). arXiv:cs.CY/1808.07261
[21]
László A Jeni, Jeffrey F Cohn, and Fernando De La Torre. 2013. Facing Imbalanced Data Recommendations for the Use of Performance Metrics. Int Conf Affect Comput Intell Interact Workshops 2013 (2013), 245--251.
[22]
Marius Kloft, Felix Stiehler, Zhilin Zheng, and Niels Pinkwart. 2014. Predicting MOOC dropout over weeks using machine learning methods. In Proc. EMNLP 2014 Workshop on Analysis of Large Scale Social Interaction in MOOCs. ACL, 60--65.
[23]
ZC Lipton, A Chouldechova, and J McAuley. 2018. Does mitigating ML's impact disparity require treatment disparity? (2018). arXiv:1711.07076
[24]
Lydia T Liu, Sarah Dean, Esther Rolf, Max Simchowitz, and Moritz Hardt. 2018. Delayed Impact of Fair Machine Learning. (March 2018). arXiv:cs.LG/1803.04383
[25]
Matthew C Makel and Jonathan A Plucker. 2014. Facts are more important than novelty: Replication in the education sciences. Educ. Res. 43, 6 (2014), 304--316.
[26]
P M Moreno-Marcos, C Alario-Hoyos, P J Muñoz-Merino, and C Delgado Kloos. 2018. Prediction in MOOCs: A review and future research directions. IEEE Trans. Learn. Technol. NA, 99 (2018), 1--1.
[27]
P Prinsloo and S Slade. 2014. Educational triage in open distance learning: Walking a moral tightrope. The Intl. Rev. of Res. in Open and Distributed Learning 15, 4 (Aug. 2014).
[28]
Lynne D Roberts, Vanessa Chang, and David Gibson. 2017. Ethical Considerations in Adopting a University- and System-Wide Approach to Data and Learning Analytics. In Big Data and Learning Analytics in Higher Education: Current Theory and Practice, Ben Kei Daniel (Ed.). Springer, Cham, 89--108.
[29]
Andrea Romei and Salvatore Ruggieri. 2014. A multidisciplinary survey on discrimination analysis. Knowl. Eng. Rev. 29, 5 (Nov. 2014), 582--638.
[30]
D Sculley, Jasper Snoek, Alex Wiltschko, and Ali Rahimi. 2018. Winner's Curse? On Pace, Progress, and Empirical Rigor. In ICLR Workshops. Vancouver, CA.
[31]
Sharon Slade and Paul Prinsloo. 2013. Learning Analytics: Ethical Issues and Dilemmas. American Behavioral Scientist 57, 10 (March 2013), 1510--1529.
[32]
C Taylor, K Veeramachaneni, and U O'Reilly. 2014. Likely to stop? Predicting Stopout in Massive Open Online Courses. (2014). arXiv:cs.CY/1408.3382
[33]
J Whitehill, K Mohan, D Seaton, Y Rosen, and D Tingley. 2017. Delving Deeper into MOOC Student Dropout Prediction. (Feb. 2017). arXiv:cs.AI/1702.06404
[34]
J Whitehill, J Williams, G Lopez, C Coleman, and J Reich. 2015. Beyond prediction: Toward automatic intervention to reduce MOOC student stopout. In Proc. EDM. 171--178.
[35]
Diyi Yang, Tanmay Sinha, David Adamson, and Carolyn Penstein Rosé. 2013. Turn on, tune in, drop out: Anticipating student dropouts in massive open online courses. In Proc. 2013 NIPS Data-driven education workshop, Vol. 11. 14.
[36]
James Zou and Londa Schiebinger. 2018. AI can be sexist and racist --- it's time to make it fair. Nature 559, 7714 (July 2018), 324.

Cited By

View all
  • (2025)Fairness for machine learning software in educationJournal of Systems and Software10.1016/j.jss.2024.112244219:COnline publication date: 1-Jan-2025
  • (2024)From biased selective labels to pseudo-labelsProceedings of the 41st International Conference on Machine Learning10.5555/3692070.3692313(6286-6324)Online publication date: 21-Jul-2024
  • (2024)Standardized interpretable fairness measures for continuous risk scoresProceedings of the 41st International Conference on Machine Learning10.5555/3692070.3692203(3327-3346)Online publication date: 21-Jul-2024
  • Show More Cited By

Index Terms

  1. Evaluating the Fairness of Predictive Student Models Through Slicing Analysis

      Recommendations

      Comments

      Information & Contributors

      Information

      Published In

      cover image ACM Other conferences
      LAK19: Proceedings of the 9th International Conference on Learning Analytics & Knowledge
      March 2019
      565 pages
      ISBN:9781450362566
      DOI:10.1145/3303772
      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

      In-Cooperation

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      Published: 04 March 2019

      Permissions

      Request permissions for this article.

      Check for updates

      Author Tags

      1. Fairness
      2. MOOCs
      3. machine learning

      Qualifiers

      • Research-article
      • Research
      • Refereed limited

      Funding Sources

      • Michigan Institute for Data Science (MIDAS)

      Conference

      LAK19

      Acceptance Rates

      Overall Acceptance Rate 236 of 782 submissions, 30%

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • Downloads (Last 12 months)202
      • Downloads (Last 6 weeks)17
      Reflects downloads up to 08 Feb 2025

      Other Metrics

      Citations

      Cited By

      View all
      • (2025)Fairness for machine learning software in educationJournal of Systems and Software10.1016/j.jss.2024.112244219:COnline publication date: 1-Jan-2025
      • (2024)From biased selective labels to pseudo-labelsProceedings of the 41st International Conference on Machine Learning10.5555/3692070.3692313(6286-6324)Online publication date: 21-Jul-2024
      • (2024)Standardized interpretable fairness measures for continuous risk scoresProceedings of the 41st International Conference on Machine Learning10.5555/3692070.3692203(3327-3346)Online publication date: 21-Jul-2024
      • (2024)Optimasi Pembelajaran: Strategi Meningkatkan Pemahaman Statistik melalui Pemahaman Kecerdasan Emosi, Spiritual, dan Intelektual MahasiswaJurnal Psikologi10.47134/pjp.v1i3.24231:3(13)Online publication date: 7-May-2024
      • (2024)Empirical Investigation of Multilayered Framework for Predicting Academic Performance in Open and Distance LearningElectronics10.3390/electronics1314280813:14(2808)Online publication date: 17-Jul-2024
      • (2024)Assessing Disparities in Predictive Modeling Outcomes for College Student Success: The Impact of Imputation Techniques on Model Performance and FairnessEducation Sciences10.3390/educsci1402013614:2(136)Online publication date: 29-Jan-2024
      • (2024)The Fairness Stitch: A Novel Approach for Neural Network DebiasingActa Informatica Pragensia10.18267/j.aip.24113:3(359-373)Online publication date: 22-Aug-2024
      • (2024)Inside the Black Box: Detecting and Mitigating Algorithmic Bias Across Racialized Groups in College Student-Success PredictionAERA Open10.1177/2332858424125874110Online publication date: 10-Jul-2024
      • (2024)Is a Seat at the Table Enough? Engaging Teachers and Students in Dataset Specification for ML in EducationProceedings of the ACM on Human-Computer Interaction10.1145/36373588:CSCW1(1-32)Online publication date: 26-Apr-2024
      • (2024)Contexts Matter but How? Course-Level Correlates of Performance and Fairness Shift in Predictive Model TransferProceedings of the 14th Learning Analytics and Knowledge Conference10.1145/3636555.3636936(713-724)Online publication date: 18-Mar-2024
      • Show More Cited By

      View Options

      Login options

      View options

      PDF

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      Figures

      Tables

      Media

      Share

      Share

      Share this Publication link

      Share on social media