Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3442188.3445934acmconferencesArticle/Chapter ViewAbstractPublication PagesfacctConference Proceedingsconference-collections
research-article
Open access

Chasing Your Long Tails: Differentially Private Prediction in Health Care Settings

Published: 01 March 2021 Publication History

Abstract

Machine learning models in health care are often deployed in settings where it is important to protect patient privacy. In such settings, methods for differentially private (DP) learning provide a general-purpose approach to learn models with privacy guarantees. Modern methods for DP learning ensure privacy through the addition of calibrated noise. The resulting privacy-preserving models are unable to learn too much information about the tails of a data distribution, resulting in a loss of accuracy that can disproportionately affect small groups. In this paper, we study the effects of DP learning in health care. We use state-of-the-art methods for DP learning to train privacy-preserving models in clinical prediction tasks, including x-ray classification of images and mortality prediction in time series data. We use these models to perform a comprehensive empirical investigation of the tradeoffs between privacy, utility, robustness to dataset shift and fairness. Our results highlight lesser-known limitations of methods for DP learning in health care, models that exhibit steep tradeoffs between privacy and utility, and models whose predictions are disproportionately influenced by large demographic groups in the training data. We discuss the costs and benefits of differentially private learning in health care with open directions for differential privacy, machine learning and health care.

Supplementary Material

suriyakumar (suriyakumar.zip)
Supplemental movie, appendix, image and software files for, Chasing Your Long Tails: Differentially Private Prediction in Health Care Settings

References

[1]
[n.d.]. MIMIC. https://mimic.physionet.org/, note = Accessed: 2020-09-30.
[2]
Martin Abadi, Andy Chu, Ian Goodfellow, H Brendan McMahan, Ilya Mironov, Kunal Talwar, and Li Zhang. 2016. Deep Learning with Differential Privacy. In Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security (Vienna, Austria) (CCS '16). ACM, New York, NY, USA, 308--318.
[3]
Accountability Act. 1996. Health insurance portability and accountability act of 1996. Public law 104 (1996), 191.
[4]
Privacy Act. 2000. Personal Information Protection and Electronic Documents Act. Department of Justice, Canada. Full text available at http://laws. justice. gc. ca/en/P-8.6/text. html (2000).
[5]
Eugene Bagdasaryan, Omid Poursaeed, and Vitaly Shmatikov. 2019. Differential privacy has disparate impact on model accuracy. In Advances in Neural Information Processing Systems. 15453--15462.
[6]
Brett K Beaulieu-Jones, William Yuan, Samuel G Finlayson, and Zhiwei Steven Wu. 2018. Privacy-preserving distributed deep learning for clinical data. arXiv preprint arXiv:1812.01484 (2018).
[7]
Cynthia J Berg, Hani K Atrash, Lisa M Koonin, and Myra Tucker. 1996. Pregnancy-related mortality in the United States, 1987--1990. Obstetrics & Gynecology 88, 2 (1996), 161--167.
[8]
Kamalika Chaudhuri, Claire Monteleoni, and Anand D Sarwate. 2011. Differentially private empirical risk minimization. Journal of Machine Learning Research 12, Mar (2011), 1069--1109.
[9]
Kamalika Chaudhuri and Staal A Vinterbo. 2013. A stability-based validation procedure for differentially private machine learning. In Advances in Neural Information Processing Systems. 2652--2660.
[10]
Zhengping Che, Sanjay Purushotham, Kyunghyun Cho, David Sontag, and Yan Liu. 2018. Recurrent Neural Networks for Multivariate Time Series with Missing Values. Scientific Reports 8, 1 (April 2018), 6085. https://doi.org/10.1038/s41598-018-24271-9
[11]
Irene Chen, Fredrik D Johansson, and David Sontag. 2018. Why is my classifier discriminatory?. In Advances in Neural Information Processing Systems. 3539--3550.
[12]
Irene Y. Chen, Emma Pierson, Sherri Rose, Shalmali Joshi, Kadija Ferryman, and Marzyeh Ghassemi. 2020. Ethical Machine Learning in Health Care. arXiv:2009.10576 [cs.CY]
[13]
Irene Y Chen, Peter Szolovits, and Marzyeh Ghassemi. 2019. Can AI Help Reduce Disparities in General Medical and Mental Health Care? AMA Journal of Ethics 21, 2 (2019), 167--179.
[14]
Xiangyi Chen, Zhiwei Steven Wu, and Mingyi Hong. 2020. Understanding Gradient Clipping in Private SGD: A Geometric Perspective. arXiv preprint arXiv:2006.15429 (2020).
[15]
Joseph Paul Cohen, Paul Morrison, Lan Dao, Karsten Roth, Tim Q Duong, and Marzyeh Ghassemi. 2020. Covid-19 image data collection: Prospective predictions are the future. arXiv preprint arXiv:2006.11988 (2020).
[16]
R Dennis Cook and Sanford Weisberg. 1980. Characterizations of an empirical influence function for detecting influential cases in regression. Technometrics 22, 4 (1980), 495--508.
[17]
David R Cox. 1972. Regression models and life-tables. Journal of the Royal Statistical Society: Series B (Methodological) 34, 2 (1972), 187--202.
[18]
Peter Croft, Douglas G Altman, Jonathan J Deeks, Kate M Dunn, Alastair D Hay, Harry Hemingway, Linda LeResche, George Peat, Pablo Perel, Steffen E Petersen, et al. 2015. The science of clinical practice: disease diagnosis or patient prognosis? Evidence about "what is likely to happen" should shape clinical practice. BMC medicine 13, 1 (2015), 20.
[19]
Rachel Cummings, Varun Gupta, Dhamma Kimpara, and Jamie Morgenstern. 2019. On the compatibility of privacy and fairness. In Adjunct Publication of the 27th Conference on User Modeling, Adaptation and Personalization. 309--315.
[20]
Thomas Davenport and Ravi Kalakota. 2019. The potential for artificial intelligence in healthcare. Future healthcare journal 6, 2 (2019), 94.
[21]
Sharon E Davis, Thomas A Lasko, Guanhua Chen, Edward D Siew, and Michael E Matheny. 2017. Calibration drift in regression and machine learning models for acute kidney injury. Journal of the American Medical Informatics Association 24, 6 (2017), 1052--1061.
[22]
Cynthia Dwork, Frank McSherry, Kobbi Nissim, and Adam Smith. 2006. Calibrating noise to sensitivity in private data analysis. In Theory of cryptography conference. Springer, 265--284.
[23]
Cynthia Dwork, Aaron Roth, et al. 2014. The algorithmic foundations of differential privacy. Foundations and Trends in Theoretical Computer Science 9, 3-4 (2014), 211--407.
[24]
Cynthia Dwork, Adam Smith, Thomas Steinke, and Jonathan Ullman. 2017. Exposed! a survey of attacks on private data. (2017).
[25]
Michael D Ekstrand, Rezvan Joshaghani, and Hoda Mehrpouyan. 2018. Privacy for all: Ensuring fair and equitable privacy protections. In Conference on Fairness, Accountability and Transparency. 35--47.
[26]
Nwamaka Denise Eneanya, Wei Yang, and Peter Philip Reese. 2019. Reconsidering the consequences of using race to estimate kidney function. Jama 322, 2 (2019), 113--114.
[27]
Tom Farrand, Fatemehsadat Mireshghallah, Sahib Singh, and Andrew Trask. 2020. Neither Private Nor Fair: Impact of Data Imbalance on Utility and Fairness in Differential Privacy. arXiv preprint arXiv:2009.06389 (2020).
[28]
Vitaly Feldman. 2020. Does learning require memorization? a short tale about a long tail. In Proceedings of the 52nd Annual ACM SIGACT Symposium on Theory of Computing. 954--959.
[29]
Vitaly Feldman and Tijana Zrnic. 2020. Individual Privacy Accounting via a Renyi Filter. arXiv preprint arXiv:2008.11193 (2020).
[30]
Kadija Ferryman and Mikaela Pitcan. 2018. Fairness in precision medicine. Data & Society (2018).
[31]
Matt Fredrikson, Somesh Jha, and Thomas Ristenpart. 2015. Model inversion attacks that exploit confidence information and basic countermeasures. In Proceedings of the 22nd ACM SIGSAC Conference on Computer and Communications Security. 1322--1333.
[32]
Quan Geng, Wei Ding, Ruiqi Guo, and Sanjiv Kumar. 2020. Tight Analysis of Privacy and Utility Tradeoff in Approximate Differential Privacy. In International Conference on Artificial Intelligence and Statistics. 89--99.
[33]
Marzyeh Ghassemi, Tristan Naumann, Peter Schulam, Andrew L Beam, and Rajesh Ranganath. 2018. Opportunities in machine learning for healthcare. arXiv preprint arXiv:1806.00388 (2018).
[34]
Brad N Greenwood, Seth Carnahan, and Laura Huang. 2018. Patient-physician gender concordance and increased mortality among female heart attack patients. Proceedings of the National Academy of Sciences 115, 34 (2018), 8569--8574.
[35]
Brad N Greenwood, Rachel R Hardeman, Laura Huang, and Aaron Sojourner. 2020. Physician-patient racial concordance and disparities in birthing mortality for newborns. Proceedings of the National Academy of Sciences 117, 35 (2020), 21194--21200.
[36]
Varun Gulshan, Lily Peng, Marc Coram, Martin C Stumpe, Derek Wu, Arunachalam Narayanaswamy, Subhashini Venugopalan, Kasumi Widner, Tom Madams, Jorge Cuadros, Ramasamy Kim, Rajiv Raman, Philip C Nelson, Jessica L Mega, and Dale R Webster. 2016. Development and Validation of a Deep Learning Algorithm for Detection of Diabetic Retinopathy in Retinal Fundus Photographs. JAMA 316, 22 (Dec. 2016), 2402--2410.
[37]
Karimollah Hajian-Tilaki. 2013. Receiver operating characteristic (ROC) curve analysis for medical diagnostic test evaluation. Caspian journal of internal medicine 4, 2 (2013), 627.
[38]
Yoni Halpern, Steven Horng, Youngduck Choi, and David Sontag. 2016. Electronic medical record phenotyping using the anchor and learn framework. Journal of the American Medical Informatics Association 23, 4 (2016), 731--740.
[39]
Han-JD. 2019. Gated Recurrent Unit with a Decay mechanism for Multivariate Time Series with Missing Values. https://github.com/Han-JD/GRU-D
[40]
Moritz Hardt, Eric Price, and Nati Srebro. 2016. Equality of opportunity in supervised learning. In Advances in neural information processing systems. 3315--3323.
[41]
Hrayr Harutyunyan, Hrant Khachatrian, David C. Kale, Greg Ver Steeg, and Aram Galstyan. 2017. Multitask Learning and Benchmarking with Clinical Time Series Data. arXiv:1703.07771 [cs, stat] (March 2017). http://arxiv.org/abs/1703. 07771 arXiv:1703.07771.
[42]
Michael B. Hawes. 2020. Implementing Differential Privacy: Seven Lessons From the 2020 United States Census. Harvard Data Science Review (30 4 2020). https://doi.org/10.1162/99608f92.353c6f99 https://hdsr.mitpress.mit.edu/pub/dgg03vo6.
[43]
Asefeh Heiat, Cary P Gross, and Harlan M Krumholz. 2002. Representation of the elderly, women, and minorities in heart failure clinical trials. Archives of internal medicine 162, 15 (2002).
[44]
Diana Herrera-Perez, Alyson Haslam, Tyler Crain, Jennifer Gill, Catherine Livingston, Victoria Kaestner, Michael Hayes, Dan Morgan, Adam S Cifu, and Vinay Prasad. 2019. Meta-Research: A comprehensive review of randomized clinical trials in three medical journals reveals 396 medical reversals. Elife 8 (2019), e45183.
[45]
G. Hripcsak, P.B. Ryan, J.D. Duke, N.H. Shah, R.W. Park, V. Huser, M.A. Suchard, M.J. Schuemie, F.J. DeFalco, A. Perotte, et al. 2016. Characterizing treatment pathways at scale using the OHDSI network. Proceedings of the National Academy of Sciences 113, 27 (2016), 7329--7336.
[46]
Matthew Jagielski, Jonathan Ullman, and Alina Oprea. 2020. Auditing Differentially Private Machine Learning: How Private is Private SGD? arXiv preprint arXiv:2006.07709 (2020).
[47]
Bargav Jayaraman and David Evans. 2019. Evaluating differentially private machine learning in practice. In 28th {USENIX} Security Symposium ({ USENIX} Security 19). 1895--1912.
[48]
Alistair E. W. Johnson, Tom J. Pollard, Lu Shen, Li-wei H. Lehman, Mengling Feng, Mohammad Ghassemi, Benjamin Moody, Peter Szolovits, Leo Anthony Celi, and Roger G. Mark. 2016. MIMIC-III, a freely accessible critical care database. Scientific Data 3, 1 (May 2016), 1--9. https://doi.org/10.1038/sdata.2016.35
[49]
James Jordon, Daniel Jarrett, Jinsung Yoon, Tavian Barnes, Paul Elbers, Patrick Thoral, Ari Ercole, Cheng Zhang, Danielle Belgrave, and Mihaela van der Schaar. 2020. Hide-and-Seek Privacy Challenge. arXiv preprint arXiv:2007.12087 (2020).
[50]
Christopher Jung, Katrina Ligett, Seth Neel, Aaron Roth, Saeed Sharifi-Malvajerdi, and Moshe Shenfeld. 2019. A New Analysis of Differential Privacy's Generalization Guarantees. arXiv preprint arXiv:1909.03577 (2019).
[51]
Kenneth Jung and Nigam H Shah. 2015. Implications of non-stationarity on predictive modeling using EHRs. Journal of biomedical informatics 58 (2015), 168--174.
[52]
Michael Kearns, Seth Neel, Aaron Roth, and Zhiwei Steven Wu. 2018. Preventing fairness gerrymandering: Auditing and learning for subgroup fairness. In International Conference on Machine Learning. 2564--2572.
[53]
Michael Kearns, Seth Neel, Aaron Roth, and Zhiwei Steven Wu. 2019. An empirical study of rich subgroup fairness for machine learning. In Proceedings of the Conference on Fairness, Accountability, and Transparency. 100--109.
[54]
Michael Kearns, Aaron Roth, Zhiwei Steven Wu, and Grigory Yaroslavtsev. 2015. Privacy for the protected (only). arXiv preprint arXiv:1506.00242 (2015).
[55]
Christopher J Kelly, Alan Karthikesalingam, Mustafa Suleyman, Greg Corrado, and Dominic King. 2019. Key challenges for delivering clinical impact with artificial intelligence. BMC medicine 17, 1 (2019), 195.
[56]
Robert Kocher, Ezekiel J Emanuel, and Nancy-Ann M DeParle. 2010. The Affordable Care Act and the future of clinical medicine: the opportunities and challenges. Annals of internal medicine 153, 8 (2010), 536--539.
[57]
Pang Wei Koh and Percy Liang. 2017. Understanding black-box predictions via influence functions. In Proceedings of the 34th International Conference on Machine Learning-Volume 70.JMLR. org, 1885--1894.
[58]
Pang Wei W Koh, Kai-Siang Ang, Hubert Teo, and Percy S Liang. 2019. On the accuracy of influence functions for measuring group effects. In Advances in Neural Information Processing Systems. 5255--5265.
[59]
Yann LeCun, Corinna Cortes, and CJ Burges. 2010. MNIST handwritten digit database. (2010).
[60]
Ninghui Li, Tiancheng Li, and Suresh Venkatasubramanian. 2007. t-closeness: Privacy beyond k-anonymity and l-diversity. In 2007 IEEE 23rd International Conference on Data Engineering. IEEE, 106--115.
[61]
Jingcheng Liu and Kunal Talwar. 2019. Private selection from private candidates. In Proceedings of the 51st Annual ACM SIGACT Symposium on Theory of Computing. 298--309.
[62]
Toni Martin. 2011. The color of kidneys. American Journal of Kidney Diseases 58, 5 (2011), A27-A28.
[63]
J Michael McGinnis, Leigh Stuckhardt, Robert Saunders, Mark Smith, et al. 2013. Best care at lower cost: the path to continuously learning health care in America. National Academies Press.
[64]
Liangyuan Na, Cong Yang, Chi-Cheng Lo, Fangyuan Zhao, Yoshimi Fukuoka, and Anil Aswani. 2018. Feasibility of reidentifying individuals in large national physical activity data sets from which protected health information has been removed with use of machine learning. JAMA network open 1, 8 (2018), e186040-e186040.
[65]
A Narayanan and V Shmatikov. 2008. Robust de-anonymization of large sparse datasets [Netflix]. In IEEE Symposium on Research in Security and Privacy, Oakland, CA.
[66]
Seth Neel, Aaron Roth, Giuseppe Vietri, and Zhiwei Steven Wu. 2019. Differentially private objective perturbation: Beyond smoothness and convexity. arXiv preprint arXiv:1909.01783 (2019).
[67]
Bret Nestor, Matthew B A McDermott, Willie Boag, Gabriela Berner, Tristan Naumann, Michael C Hughes, Anna Goldenberg, and Marzyeh Ghassemi. 2019. Feature Robustness in Non-stationary Health Records: Caveats to Deployable Model Performance in Common Clinical Machine Learning Tasks.
[68]
Thông T Nguyën, Xiaokui Xiao, Yin Yang, Siu Cheung Hui, Hyejin Shin, and Junbum Shin. 2016. Collecting and analyzing data from smart device users with local differential privacy. arXiv preprint arXiv:1606.05053 (2016).
[69]
Kobbi Nissim and Uri Stemmer. 2015. On the generalization properties of differential privacy. CoRR, abs/1504.05800 (2015).
[70]
Ziad Obermeyer, Brian Powers, Christine Vogeli, and Sendhil Mullainathan. 2019. Dissecting racial bias in an algorithm used to manage the health of populations. Science 366, 6464 (2019), 447--453.
[71]
K.J. O'malley, K.F. Cook, M.D. Price, K.R. Wildes, J.F. Hurdle, and C.M. Ashton. 2005. Measuring diagnoses: ICD code accuracy. Health Services Research 40, 5p2 (2005), 1620--1639.
[72]
Jennifer M Orsi, Helen Margellos-Anast, and Steven Whitman. 2010. Black-white health disparities in the United States and Chicago: a 15-year progress analysis. American journal of public health 100, 2 (2010), 349--356.
[73]
Nicolas Papernot, Martín Abadi, Úlfar Erlingsson, Ian Goodfellow, and Kunal Talwar. 2016. Semi-supervised Knowledge Transfer for Deep Learning from Private Training Data. (Oct. 2016). arXiv:1610.05755 [stat.ML]
[74]
Nicolas Papernot, Steve Chien, Shuang Song, Abhradeep Thakurta, and Ulfar Erlingsson. 2020. Making the Shoe Fit: Architectures, Initializations, and Tuning for Learning with Privacy. https://openreview.net/forum?id=rJg851rYwH
[75]
Fabian Pedregosa, Gaël Varoquaux, Alexandre Gramfort, Vincent Michel, Bertrand Thirion, Olivier Grisel, Mathieu Blondel, Peter Prettenhofer, Ron Weiss, Vincent Dubourg, Jake Vanderplas, Alexandre Passos, David Cournapeau, Matthieu Brucher, Matthieu Perrot, and Édouard Duchesnay. 2011. Scikit-learn: Machine Learning in Python. Journal of Machine Learning Research 12 (Oct. 2011), 2825--2830. http://jmlr.csail.mit.edu/papers/v12/pedregosa11a.html
[76]
Stephen R Pfohl, Andrew M Dai, and Katherine Heller. 2019. Federated and Differentially Private Learning for Electronic Health Records. arXiv preprint arXiv:1911.05861 (2019).
[77]
Stephen R Pfohl, Agata Foryciarz, and Nigam H Shah. 2020. An Empirical Characterization of Fair Machine Learning For Clinical Risk Prediction. arXiv preprint arXiv:2007.10306 (2020).
[78]
Stephan Rabanser, Stephan Günnemann, and Zachary Lipton. 2019. Failing loudly: an empirical study of methods for detecting dataset shift. In Advances in Neural Information Processing Systems. 1394--1406.
[79]
Alvin Rajkomar, Michaela Hardt, Michael D Howell, Greg Corrado, and Marshall H Chin. 2018. Ensuring fairness in machine learning to advance health equity. Annals of internal medicine 169, 12 (2018), 866--872.
[80]
Alvin Rajkomar, Eyal Oren, Kai Chen, Andrew M Dai, Nissan Hajaj, Michaela Hardt, Peter J Liu, Xiaobing Liu, Jake Marcus, Mimi Sun, et al. 2018. Scalable and accurate deep learning with electronic health records. NPJ Digital Medicine 1, 1 (2018), 18.
[81]
Sherri Rose. 2018. Machine learning for prediction in electronic health data. JAMA network open 1, 4 (2018), e181404-e181404.
[82]
Christopher G Schwarz, Walter K Kremers, Terry M Therneau, Richard R Sharp, Jeffrey L Gunter, Prashanthi Vemuri, Arvin Arani, Anthony J Spychalla, Kejal Kantarci, David S Knopman, et al. 2019. Identification of anonymous MRI research participants with face-recognition software. New England Journal of Medicine 381, 17 (2019), 1684--1686.
[83]
Laleh Seyyed-Kalantari, Guanxiong Liu, Matthew McDermott, and Marzyeh Ghassemi. 2020. CheXclusion: Fairness gaps in deep chest X-ray classifiers. arXiv preprint arXiv:2003.00827 (2020).
[84]
Reza Shokri, Marco Stronati, Congzheng Song, and Vitaly Shmatikov. 2017. Membership inference attacks against machine learning models. In 2017 IEEE Symposium on Security and Privacy (SP). IEEE, 3--18.
[85]
Shuang Song, Om Thakkar, and Abhradeep Thakurta. 2020. Characterizing private clipped gradient descent on convex generalized linear problems. arXiv preprint arXiv:2006.06783 (2020).
[86]
Stephen M Stigler. 1989. Francis Galton's account of the invention of correlation. Statist. Sci. (1989), 73--79.
[87]
Karien Stronks, Nicolien F Wieringa, and Anita Hardon. 2013. Confronting diversity in the production of clinical evidence goes beyond merely including under-represented groups in clinical trials. Trials 14, 1 (2013), 1--6.
[88]
Adarsh Subbaswamy, Peter Schulam, and Suchi Saria. 2018. Preventing failures due to dataset shift: Learning predictive models that transport. arXiv preprint arXiv:1812.04597 (2018).
[89]
Latanya Sweeney. 2002. k-anonymity: A model for protecting privacy. International Journal of Uncertainty, Fuzziness and Knowledge-Based Systems 10, 05 (2002), 557--570.
[90]
Latanya Sweeney. 2015. Only you, your doctor, and many others may know. Technology Science 2015092903, 9 (2015), 29.
[91]
Jun Tang, Aleksandra Korolova, Xiaolong Bai, Xueqiang Wang, and Xiaofeng Wang. 2017. Privacy loss in apple's implementation of differential privacy on macos 10.12. arXiv preprint arXiv:1709.02753 (2017).
[92]
Nenad Tomašev, Xavier Glorot, Jack W Rae, Michal Zielinski, Harry Askham, Andre Saraiva, Anne Mottram, Clemens Meyer, Suman Ravuri, Ivan Protsyuk, Alistair Connell, Cían O Hughes, Alan Karthikesalingam, Julien Cornebise, Hugh Montgomery, Geraint Rees, Chris Laing, Clifton R Baker, Kelly Peterson, Ruth Reeves, Demis Hassabis, Dominic King, Mustafa Suleyman, Trevor Back, Christopher Nielson, Joseph R Ledsam, and Shakir Mohamed. 2019. A clinically applicable approach to continuous prediction of future acute kidney injury. Nature 572, 7767 (Aug. 2019), 116--119.
[93]
Eric J Topol. 2019. High-performance medicine: the convergence of human and artificial intelligence. Nature medicine 25, 1 (2019), 44--56.
[94]
Justin Travers, Suzanne Marsh, Mathew Williams, Mark Weatherall, Brent Caldwell, Philippa Shirtcliffe, Sarah Aldington, and Richard Beasley. 2007. External validity of randomised controlled trials in asthma: to whom do the results of the trials apply? Thorax 62, 3 (2007), 219--223.
[95]
Salil Vadhan. 2017. The complexity of differential privacy. In Tutorials on the Foundations of Cryptography. Springer, 347--450.
[96]
Giuseppe Vietri, Borja Balle, Akshay Krishnamurthy, and Zhiwei Steven Wu. 2020. Private Reinforcement Learning with PAC and Regret Guarantees. arXiv preprint arXiv:2009.09052 (2020).
[97]
Darshali A Vyas, Leo G Eisenstein, and David S Jones. 2020. Hidden in Plain Sight---Reconsidering the Use of Race Correction in Clinical Algorithms.
[98]
Shirly Wang, Matthew B. A. McDermott, Geeticka Chauhan, Michael C. Hughes, Tristan Naumann, and Marzyeh Ghassemi. 2019. MIMIC-Extract: A Data Extraction, Preprocessing, and Representation Pipeline for MIMIC-III. arXiv:1907.08322 [cs, stat] (July 2019). http://arxiv.org/abs/1907.08322 arXiv:1907.08322.
[99]
Xiaosong Wang, Yifan Peng, Le Lu, Zhiyong Lu, Mohammadhadi Bagheri, and Ronald M Summers. 2017. Chestx-ray8: Hospital-scale chest x-ray database and benchmarks on weakly-supervised classification and localization of common thorax diseases. In Proceedings of the IEEE conference on computer vision and pattern recognition. 2097--2106.
[100]
Denny Wu, Hirofumi Kobayashi, Charles Ding, Lei Cheng, and Keisuke Goda Marzyeh Ghassemi. 2019. Modeling the Biological Pathology Continuum with HSIC-regularized Wasserstein Auto-encoders. arXiv:1901.06618 [cs, stat] (Jan. 2019). http://arxiv.org/abs/1901.06618 arXiv:1901.06618.
[101]
Xi Wu, Fengan Li, Arun Kumar, Kamalika Chaudhuri, Somesh Jha, and Jeffrey Naughton. 2017. Bolt-on differential privacy for scalable stochastic gradient descent-based analytics. In Proceedings of the 2017 ACM International Conference on Management of Data. 1307--1322.
[102]
Han Xiao, Kashif Rasul, and Roland Vollgraf. 2017. Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms. arXiv preprint arXiv:1708.07747 (2017).
[103]
S. Yu, Y. Ma, J. Gronsbell, T. Cai, A.N. Ananthakrishnan, V.S. Gainer, S.E. Churchill, P. Szolovits, S.N. Murphy, I.S. Kohane, et al. 2017. Enabling phenotypic big data with PheNorm. Journal of the American Medical Informatics Association 25, 1 (2017), 54--60.
[104]
Chiyuan Zhang, Samy Bengio, Moritz Hardt, Benjamin Recht, and Oriol Vinyals. 2016. Understanding deep learning requires rethinking generalization. arXiv preprint arXiv:1611.03530 (2016).

Cited By

View all
  • (2024)Privacy-Enhancing Technologies in Biomedical Data ScienceAnnual Review of Biomedical Data Science10.1146/annurev-biodatasci-120423-1201077:1(317-343)Online publication date: 23-Aug-2024
  • (2024)The unfair side of Privacy Enhancing Technologies: addressing the trade-offs between PETs and fairnessProceedings of the 2024 ACM Conference on Fairness, Accountability, and Transparency10.1145/3630106.3659024(2047-2059)Online publication date: 3-Jun-2024
  • (2024)SoK: Unintended Interactions among Machine Learning Defenses and Risks2024 IEEE Symposium on Security and Privacy (SP)10.1109/SP54263.2024.00243(2996-3014)Online publication date: 19-May-2024
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
FAccT '21: Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency
March 2021
899 pages
ISBN:9781450383097
DOI:10.1145/3442188
This work is licensed under a Creative Commons Attribution International 4.0 License.

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 01 March 2021

Check for updates

Author Tags

  1. fairness
  2. health care
  3. machine learning
  4. privacy
  5. robustness

Qualifiers

  • Research-article
  • Research
  • Refereed limited

Conference

FAccT '21
Sponsor:

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)391
  • Downloads (Last 6 weeks)48
Reflects downloads up to 13 Jan 2025

Other Metrics

Citations

Cited By

View all
  • (2024)Privacy-Enhancing Technologies in Biomedical Data ScienceAnnual Review of Biomedical Data Science10.1146/annurev-biodatasci-120423-1201077:1(317-343)Online publication date: 23-Aug-2024
  • (2024)The unfair side of Privacy Enhancing Technologies: addressing the trade-offs between PETs and fairnessProceedings of the 2024 ACM Conference on Fairness, Accountability, and Transparency10.1145/3630106.3659024(2047-2059)Online publication date: 3-Jun-2024
  • (2024)SoK: Unintended Interactions among Machine Learning Defenses and Risks2024 IEEE Symposium on Security and Privacy (SP)10.1109/SP54263.2024.00243(2996-3014)Online publication date: 19-May-2024
  • (2024)Differentially private and explainable boosting machine with enhanced utilityNeurocomputing10.1016/j.neucom.2024.128424(128424)Online publication date: Aug-2024
  • (2024)A comprehensive survey and taxonomy on privacy-preserving deep learningNeurocomputing10.1016/j.neucom.2024.127345576(127345)Online publication date: Apr-2024
  • (2024)Artificial Intelligence and Machine Learning in Precision Health: An Overview of Methods, Challenges, and Future DirectionsDynamics of Disasters10.1007/978-3-031-74006-0_2(15-53)Online publication date: 24-Dec-2024
  • (2023)Taking Off with AI: Lessons from Aviation for HealthcareProceedings of the 3rd ACM Conference on Equity and Access in Algorithms, Mechanisms, and Optimization10.1145/3617694.3623224(1-14)Online publication date: 30-Oct-2023
  • (2023)Private Graph Data Release: A SurveyACM Computing Surveys10.1145/356908555:11(1-39)Online publication date: 22-Feb-2023
  • (2023)Learning Privacy-Preserving Student Networks via Discriminative-Generative DistillationIEEE Transactions on Image Processing10.1109/TIP.2022.322641632(116-127)Online publication date: 2023
  • (2023)Presentation matters for AI-generated clinical adviceNature Human Behaviour10.1038/s41562-023-01721-77:11(1833-1835)Online publication date: 20-Nov-2023
  • Show More Cited By

View Options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Login options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media