Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3133956.3134097acmconferencesArticle/Chapter ViewAbstractPublication PagesccsConference Proceedingsconference-collections
research-article
Public Access

Use Privacy in Data-Driven Systems: Theory and Experiments with Machine Learnt Programs

Published: 30 October 2017 Publication History

Abstract

This paper presents an approach to formalizing and enforcing a class of use privacy properties in data-driven systems. In contrast to prior work, we focus on use restrictions on proxies (i.e. strong predictors) of protected information types. Our definition relates proxy use to intermediate computations that occur in a program, and identify two essential properties that characterize this behavior: 1) its result is strongly associated with the protected information type in question, and 2) it is likely to causally affect the final output of the program. For a specific instantiation of this definition, we present a program analysis technique that detects instances of proxy use in a model, and provides a witness that identifies which parts of the corresponding program exhibit the behavior. Recognizing that not all instances of proxy use of a protected information type are inappropriate, we make use of a normative judgment oracle that makes this inappropriateness determination for a given witness. Our repair algorithm uses the witness of an inappropriate proxy use to transform the model into one that provably does not exhibit proxy use, while avoiding changes that unduly affect classification accuracy. Using a corpus of social datasets, our evaluation shows that these algorithms are able to detect proxy use instances that would be difficult to find using existing techniques, and subsequently remove them while maintaining acceptable classification performance.

References

[1]
Paul Barford, Igor Canadi, Darja Krushevskaja, Qiang Ma, and S. Muthukrishnan 2014. Adscape: Harvesting and Analyzing Online Display Ads Proceedings of the 23rd International Conference on World Wide Web. International World Wide Web Conferences Steering Committee, Republic and Canton of Geneva, Switzerland, 597--608.
[2]
Raef Bassily, Adam Smith, and Abhradeep Thakurta 2014. Private Empirical Risk Minimization: Efficient Algorithms and Tight Error Bounds 55th IEEE Annual Symposium on Foundations of Computer Science, FOCS 2014, Philadelphia, PA, USA, October 18--21, 2014. 464--473.
[3]
Richard Berk and Justin Bleich 2014. Forecasts of Violence to Inform Sentencing Decisions. Journal of Quantitative Criminology Vol. 30, 1 (2014), 79--96. .acm.org/10.1145/2381966.2381969
[4]
Rich Zemel, Yu Wu, Kevin Swersky, Toni Pitassi, and Cynthia Dwork 2013. Learning Fair Representations. In Proceedings of the Internetional Conference on Machine Learning. endthebibliography

Cited By

View all
  • (2024)Goal Orientation for Fair Machine Learning AlgorithmsProduction and Operations Management10.1177/10591478241234998Online publication date: 18-Mar-2024
  • (2024)MirrorFair: Fixing Fairness Bugs in Machine Learning Software via Counterfactual PredictionsProceedings of the ACM on Software Engineering10.1145/36608011:FSE(2121-2143)Online publication date: 12-Jul-2024
  • (2024)Approaching the Information-Theoretic Limit of Privacy Disclosure With Utility GuaranteesIEEE Transactions on Information Forensics and Security10.1109/TIFS.2024.335441219(3339-3352)Online publication date: 1-Jan-2024
  • Show More Cited By

Index Terms

  1. Use Privacy in Data-Driven Systems: Theory and Experiments with Machine Learnt Programs

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    CCS '17: Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security
    October 2017
    2682 pages
    ISBN:9781450349468
    DOI:10.1145/3133956
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 30 October 2017

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. causal analysis
    2. privacy
    3. proxy
    4. use privacy

    Qualifiers

    • Research-article

    Funding Sources

    Conference

    CCS '17
    Sponsor:

    Acceptance Rates

    CCS '17 Paper Acceptance Rate 151 of 836 submissions, 18%;
    Overall Acceptance Rate 1,261 of 6,999 submissions, 18%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)125
    • Downloads (Last 6 weeks)13
    Reflects downloads up to 15 Oct 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)Goal Orientation for Fair Machine Learning AlgorithmsProduction and Operations Management10.1177/10591478241234998Online publication date: 18-Mar-2024
    • (2024)MirrorFair: Fixing Fairness Bugs in Machine Learning Software via Counterfactual PredictionsProceedings of the ACM on Software Engineering10.1145/36608011:FSE(2121-2143)Online publication date: 12-Jul-2024
    • (2024)Approaching the Information-Theoretic Limit of Privacy Disclosure With Utility GuaranteesIEEE Transactions on Information Forensics and Security10.1109/TIFS.2024.335441219(3339-3352)Online publication date: 1-Jan-2024
    • (2024)Attack Risk Analysis in Data Anonymization in Internet of ThingsIEEE Transactions on Computational Social Systems10.1109/TCSS.2023.324308911:4(4986-4993)Online publication date: Aug-2024
    • (2024)Discrimination for the sake of fairness by design and its legal frameworkComputer Law & Security Review10.1016/j.clsr.2023.10591652(105916)Online publication date: Apr-2024
    • (2024)MBFair: a model-based verification methodology for detecting violations of individual fairnessSoftware and Systems Modeling10.1007/s10270-024-01184-yOnline publication date: 10-Jun-2024
    • (2023)Fair densities via boosting the sufficient statistics of exponential familiesProceedings of the 40th International Conference on Machine Learning10.5555/3618408.3619739(32105-32144)Online publication date: 23-Jul-2023
    • (2023)A Review of Partial Information Decomposition in Algorithmic Fairness and ExplainabilityEntropy10.3390/e2505079525:5(795)Online publication date: 13-May-2023
    • (2023)Algorithmische Fairness in der polizeilichen Ermittlungsarbeit:Ethische Analyse von Verfahren des maschinellen Lernens zur GesichtserkennungTATuP - Zeitschrift für Technikfolgenabschätzung in Theorie und Praxis10.14512/tatup.32.1.2432:1(24-29)Online publication date: 23-Mar-2023
    • (2023)Disambiguating Algorithmic Bias: From Neutrality to JusticeProceedings of the 2023 AAAI/ACM Conference on AI, Ethics, and Society10.1145/3600211.3604695(691-704)Online publication date: 8-Aug-2023
    • Show More Cited By

    View Options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Get Access

    Login options

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media