Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3375627.3375866acmconferencesArticle/Chapter ViewAbstractPublication PagesaiesConference Proceedingsconference-collections
research-article
Open access

Why Reliabilism Is not Enough: Epistemic and Moral Justification in Machine Learning

Published: 07 February 2020 Publication History
  • Get Citation Alerts
  • Abstract

    In this paper we argue that standard calls for explainability that focus on the epistemic inscrutability of black-box machine learning models may be misplaced. If we presume, for the sake of this paper, that machine learning can be a source of knowledge, then it makes sense to wonder what kind of \em justification it involves. How do we rationalize on the one hand the seeming justificatory black box with the observed wide adoption of machine learning? We argue that, in general, people implicitly adoptreliabilism regarding machine learning. Reliabilism is an epistemological theory of epistemic justification according to which a belief is warranted if it has been produced by a reliable process or method \citegoldman2012reliabilism. We argue that, in cases where model deployments require \em moral justification, reliabilism is not sufficient, and instead justifying deployment requires establishing robust human processes as a moral "wrapper'' around machine outputs. We then suggest that, in certain high-stakes domains with moral consequences, reliabilism does not provide another kind of necessary justification---moral justification. Finally, we offer cautions relevant to the (implicit or explicit) adoption of the reliabilist interpretation of machine learning.

    References

    [1]
    Kai Arulkumaran, Antoine Cully, and Julian Togelius. 2019. Alphastar: An evolutionary computation perspective. arXiv preprint arXiv:1902.01724 (2019).
    [2]
    Muzammil Bashir. 2019. Deep Learning Approach to Trespass Detection using Video Surveillance Data. (2019).
    [3]
    Ruha Benjamin. 2019. Race after technology: Abolitionist tools for the new jim code. John Wiley & Sons.
    [4]
    Richard Berk. 2019. Machine learning risk assessments in criminal justice settings. Springer.
    [5]
    Reuben Binns, Max Van Kleek, Michael Veale, Ulrik Lyngs, Jun Zhao, and Nigel Shadbolt. 2018. 'It's Reducing a Human Being to a Percentage': Perceptions of Justice in Algorithmic Decisions. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems. ACM, 377.
    [6]
    Joanna J Bryson, Mihailis E Diamantis, and Thomas D Grant. 2017. Of, for, and by the people: the legal lacuna of synthetic persons. Artificial Intelligence and Law 25, 3 (2017), 273--291.
    [7]
    Jenna Burrell. 2016. Howthe machine "thinks": Understanding opacity in machine learning algorithms. Big Data & Society 3, 1 (2016), 2053951715622512.
    [8]
    Richmond Campbell. 2015. Moral Epistemology. In The Stanford Encyclopedia of Philosophy (winter 2015 ed.), Edward N. Zalta (Ed.). Metaphysics Research Lab, Stanford University.
    [9]
    Alexandra Chouldechova. 2017. Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. Big data 5, 2 (2017), 153--163.
    [10]
    Mark Coeckelbergh. 2019. Artificial Intelligence, Responsibility Attribution, and a Relational Justification of Explainability. Science and engineering ethics (2019), 1--18.
    [11]
    Earl Conee and Richard Feldman. 1998. The generality problem for reliabilism. Philosophical Studies 89, 1 (1998), 1--29.
    [12]
    Finale Doshi-Velez and Been Kim. 2017. Towards a rigorous science of interpretable machine learning. arXiv preprint arXiv:1702.08608 (2017).
    [13]
    Laurel Eckhouse, Kristian Lum, Cynthia Conti-Cook, and Julie Ciccolini. 2019. Layers of bias: A unified approach for understanding problems with risk assessment. Criminal Justice and Behavior 46, 2 (2019), 185--209.
    [14]
    Andre Esteva, Alexandre Robicquet, Bharath Ramsundar, Volodymyr Kuleshov, Mark DePristo, Katherine Chou, Claire Cui, Greg Corrado, Sebastian Thrun, and Jeff Dean. 2019. A guide to deep learning in healthcare. Nature medicine 25, 1 (2019), 24--29.
    [15]
    Richard Feldman. 2003. Epistemology. (2003).
    [16]
    Edmund Gettier. 1963. Is justified true belief knowledge? 1963 (1963), 273--274.
    [17]
    Alvin Goldman and Bob Beddor. 2016. Reliabilist Epistemology. In The Stanford Encyclopedia of Philosophy (winter 2016 ed.), Edward N. Zalta (Ed.). Metaphysics Research Lab, Stanford University.
    [18]
    Alvin I Goldman. 2012. Reliabilism and contemporary epistemology: essays. Oxford University Press.
    [19]
    David Gunning. 2017. Explainable artificial intelligence (xai). Defense Advanced Research Projects Agency (DARPA), nd Web 2 (2017).
    [20]
    Lars Hall, Petter Johansson, Betty Tärning, Sverker Sikström, and Thérèse Deutgen. 2010. Magic at the marketplace: Choice blindness for the taste of jam and the smell of tea. Cognition 117, 1 (2010), 54--61.
    [21]
    Andreas Holzinger, Chris Biemann, Constantinos S Pattichis, and Douglas B Kell. 2017. What do we need to build explainable AI systems for the medical domain? arXiv preprint arXiv:1712.09923 (2017).
    [22]
    Gábor Horváth, Balázs Bernáth, and Gergely Molnár. 1998. Dragonflies find crude oil visually more attractive than water: multiple-choice experiments on dragonfly polarotaxis. Naturwissenschaften 85, 6 (1998), 292--297.
    [23]
    Jochen Kruppa, Alexandra Schwarz, Gerhard Arminger, and Andreas Ziegler. 2013. Consumer credit risk: Individual probability estimates using machine learning. Expert Systems with Applications 40, 13 (2013), 5125--5131.
    [24]
    Jarrett Leplin. 2007. In defense of reliabilism. Philosophical Studies 134, 1 (2007), 31--42.
    [25]
    David Lewis. 1996. Elusive knowledge. Australasian journal of Philosophy 74, 4 (1996), 549--567.
    [26]
    Zachary C Lipton, Yu-Xiang Wang, and Alex Smola. 2018. Detecting and correcting for label shift with black box predictors. arXiv preprint arXiv:1802.03916 (2018).
    [27]
    Lydia T Liu, Sarah Dean, Esther Rolf, Max Simchowitz, and Moritz Hardt. 2018. Delayed impact of fair machine learning. arXiv preprint arXiv:1803.04383 (2018).
    [28]
    Spyros Makridakis. 2017. The forthcoming Artificial Intelligence (AI) revolution: Its impact on society and firms. Futures 90 (2017), 46--60.
    [29]
    Gary Marcus. 2018. Deep learning: A critical appraisal. arXiv preprint arXiv:1801.00631 (2018).
    [30]
    Deirdre K Mulligan, Daniel Kluttz, and Nitin Kohli. 2018. Contestability and Professionals: From Explanations to Engagement with Algorithmic Systems. Available at SSRN 3311894 (2018).
    [31]
    Harold Noonan. 2002. Routledge philosophy guidebook to Hume on knowledge. Routledge.
    [32]
    William J Prior. 2016. Virtue and knowledge: An Introduction to ancient Greek ethics. Routledge.
    [33]
    Stathis Psillos. 2007. Philosophy of science AZ. Edinburgh University Press.
    [34]
    Stathis Psillos. 2014. Causation and explanation. Routledge.
    [35]
    Willard Van Orman Quine and Joseph Silbert Ullian. 1978. The web of belief. Vol. 2. Random House New York.
    [36]
    Conrad Sachweh. 2018. General Data Protection Regulation and Explainable Machine Learning Challenges. (2018).
    [37]
    Wojciech Samek, Thomas Wiegand, and Klaus-Robert Müller. 2017. Explainable artificial intelligence: Understanding, visualizing and interpreting deep learning models. arXiv preprint arXiv:1708.08296 (2017).
    [38]
    Andrew D Selbst and Solon Barocas. 2018. The intuitive appeal of explainable machines. Fordham L. Rev. 87 (2018), 1085.
    [39]
    Andrew D Selbst, Danah Boyd, Sorelle A Friedler, Suresh Venkatasubramanian, and Janet Vertesi. 2019. Fairness and abstraction in sociotechnical systems. In Proceedings of the Conference on Fairness, Accountability, and Transparency. ACM, 59--68.
    [40]
    Stefan Thaler, Vlado Menkovski, and Milan Petkovic. 2018. Deep Learning in Information Security. arXiv preprint arXiv:1809.04332 (2018).
    [41]
    Jeremy Waldron. 2009. Judges as moral reasoners. International Journal of Constitutional Law 7, 1 (2009), 2--24.
    [42]
    Douglas Walton. 2010. Appeal to expert opinion: Arguments from authority. Penn State Press.

    Cited By

    View all
    • (2023)Knowledge representation and acquisition for ethical AI: challenges and opportunitiesEthics and Information Technology10.1007/s10676-023-09692-z25:1Online publication date: 11-Mar-2023
    • (2022)Relative explainability and double standards in medical decision-makingEthics and Information Technology10.1007/s10676-022-09646-x24:2Online publication date: 5-Apr-2022
    • (2021)Towards Accountability for Machine Learning DatasetsProceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency10.1145/3442188.3445918(560-575)Online publication date: 3-Mar-2021

    Index Terms

    1. Why Reliabilism Is not Enough: Epistemic and Moral Justification in Machine Learning

      Recommendations

      Comments

      Information & Contributors

      Information

      Published In

      cover image ACM Conferences
      AIES '20: Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society
      February 2020
      439 pages
      ISBN:9781450371100
      DOI:10.1145/3375627
      This work is licensed under a Creative Commons Attribution International 4.0 License.

      Sponsors

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      Published: 07 February 2020

      Permissions

      Request permissions for this article.

      Check for updates

      Author Tags

      1. epistemology
      2. explainability
      3. interpretability
      4. machine learning
      5. moral justification
      6. neural networks

      Qualifiers

      • Research-article

      Conference

      AIES '20
      Sponsor:

      Acceptance Rates

      Overall Acceptance Rate 61 of 162 submissions, 38%

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • Downloads (Last 12 months)226
      • Downloads (Last 6 weeks)19
      Reflects downloads up to 26 Jul 2024

      Other Metrics

      Citations

      Cited By

      View all
      • (2023)Knowledge representation and acquisition for ethical AI: challenges and opportunitiesEthics and Information Technology10.1007/s10676-023-09692-z25:1Online publication date: 11-Mar-2023
      • (2022)Relative explainability and double standards in medical decision-makingEthics and Information Technology10.1007/s10676-022-09646-x24:2Online publication date: 5-Apr-2022
      • (2021)Towards Accountability for Machine Learning DatasetsProceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency10.1145/3442188.3445918(560-575)Online publication date: 3-Mar-2021

      View Options

      View options

      PDF

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      Get Access

      Login options

      Media

      Figures

      Other

      Tables

      Share

      Share

      Share this Publication link

      Share on social media