Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3328519.3329126acmconferencesArticle/Chapter ViewAbstractPublication PagesmodConference Proceedingsconference-collections
research-article

Learning to Validate the Predictions of Black Box Machine Learning Models on Unseen Data

Published: 05 July 2019 Publication History
  • Get Citation Alerts
  • Abstract

    When end users apply a machine learning (ML) model on new unlabeled data, it is difficult for them to decide whether they can trust its predictions. Errors or shifts in the target data can lead to hard-to-detect drops in the predictive quality of the model. We therefore propose an approach to assist non-ML experts working with pretrained ML models. Our approach estimates the change in prediction performance of a model on unseen target data. It does not require explicit distributional assumptions on the dataset shift between the training and target data. Instead, a domain expert can declaratively specify typical cases of dataset shift that she expects to observe in real-world data. Based on this information, we learn a performance predictor for pretrained black box models, which can be combined with the model, and automatically warns end users in case of unexpected performance drops. We demonstrate the effectiveness of our approach on two models -- logistic regression and a neural network, applied to several real-world datasets.

    References

    [1]
    Denis Baylor, Eric Breck, Heng-Tze Cheng, and Martin Zinkevich. 2017. TFX: A TensorFlow-Based Production-Scale Machine Learning Platform. KDD (2017), 1387--1395.
    [2]
    Zachary C Lip ton, Yu-Xiang Wang, and Alex Smola. 2018. Detecting and Correcting for Label Shift with Black Box Predictors. arXiv preprint arXiv:1802.03916 (2018).
    [3]
    Stephan Rabanser, Stephan Günnemann, and Zachary C Lipton. 2018. Failing Loudly: An Empirical Study of Methods for Detecting Dataset Shift. arXiv preprint arXiv:1810.11953 (2018).
    [4]
    Sebastian Schelter, Felix Biessmann, Tim Januschowski, David Salinas, Stephan Seufert, and Gyuri Szarvas. 2018. On Challenges in Machine Learning Model Management. IEEE Data Engineering Bulletin (12 2018).
    [5]
    Sebastian Schelter, Dustin Lange, Philipp Schmidt, Meltem Celikel, Felix Biessmann, and Andreas Grafberger. 2018. Automating large-scale data quality verification. PVLDB 11, 12 (2018), 1781--1794.
    [6]
    Masashi Sugiyama, Neil D Lawrence, Anton Schwaighofer, et al. 2017. Dataset shift in machine learning. The MIT Press.

    Cited By

    View all
    • (2022)Active surrogate estimatorsProceedings of the 36th International Conference on Neural Information Processing Systems10.5555/3600270.3602053(24557-24570)Online publication date: 28-Nov-2022
    • (2022)Performance Prediction Under Dataset Shift2022 26th International Conference on Pattern Recognition (ICPR)10.1109/ICPR56361.2022.9956676(2466-2474)Online publication date: 21-Aug-2022
    • (2021)Use of Bi-Temporal ALS Point Clouds for Tree Removal Detection on Private Property in Racibórz, PolandRemote Sensing10.3390/rs1304076713:4(767)Online publication date: 19-Feb-2021
    • Show More Cited By

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    HILDA '19: Proceedings of the Workshop on Human-In-the-Loop Data Analytics
    July 2019
    67 pages
    ISBN:9781450367912
    DOI:10.1145/3328519
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 05 July 2019

    Permissions

    Request permissions for this article.

    Check for updates

    Qualifiers

    • Research-article
    • Research
    • Refereed limited

    Funding Sources

    Conference

    SIGMOD/PODS '19
    Sponsor:

    Acceptance Rates

    HILDA '19 Paper Acceptance Rate 12 of 24 submissions, 50%;
    Overall Acceptance Rate 28 of 56 submissions, 50%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)41
    • Downloads (Last 6 weeks)2

    Other Metrics

    Citations

    Cited By

    View all
    • (2022)Active surrogate estimatorsProceedings of the 36th International Conference on Neural Information Processing Systems10.5555/3600270.3602053(24557-24570)Online publication date: 28-Nov-2022
    • (2022)Performance Prediction Under Dataset Shift2022 26th International Conference on Pattern Recognition (ICPR)10.1109/ICPR56361.2022.9956676(2466-2474)Online publication date: 21-Aug-2022
    • (2021)Use of Bi-Temporal ALS Point Clouds for Tree Removal Detection on Private Property in Racibórz, PolandRemote Sensing10.3390/rs1304076713:4(767)Online publication date: 19-Feb-2021
    • (2020)Learning to Validate the Predictions of Black Box Classifiers on Unseen DataProceedings of the 2020 ACM SIGMOD International Conference on Management of Data10.1145/3318464.3380604(1289-1299)Online publication date: 11-Jun-2020

    View Options

    Get Access

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media