Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
article
Free access

A Survey on the Explainability of Supervised Machine Learning

Published: 01 May 2021 Publication History

Abstract

Predictions obtained by, e.g., artificial neural networks have a high accuracy but humans often perceive the models as black boxes. Insights about the decision making are mostly opaque for humans. Particularly understanding the decision making in highly sensitive areas such as healthcare or finance, is of paramount importance. The decision-making behind the black boxes requires it to be more transparent, accountable, and understandable for humans. This survey paper provides essential definitions, an overview of the different principles and methodologies of explainable Supervised Machine Learning (SML). We conduct a state-of-the-art survey that reviews past and recent explainable SML approaches and classifies them according to the introduced definitions. Finally, we illustrate principles by means of an explanatory case study and discuss important future directions.

References

[1]
Abdollahi, B. &#38; Nasraoui, O. (2016). Explainable restricted boltzmann machines for collaborative filtering. <i>arXiv preprint arXiv:1606.07129</i>.
[2]
Abdollahi, B. &#38; Nasraoui, O. (2017). Using explainability for constrained matrix factorization. In <i>Proceedings of the Eleventh ACM Conference on Recommender Systems</i> (pp. 79-83).
[3]
ACM (2017). Statement on algorithmic transparency and accountability.
[4]
Adadi, A. &#38; Berrada, M. (2018). Peeking inside the black-box: A survey on explainable artificial intelligence (xai). <i>IEEE Access</i>.
[5]
Adler, P., Falk, C., Friedler, S., Rybeck, G., Scheidegger, C., Smith, B., &#38; Venkatasubramanian, S. (2016). Auditing black-box models for indirect inuence. In <i>Data Mining (ICDM), 2016 IEEE 16th International Conference on</i>: IEEE.
[6]
Amatriain, X. (2017). More data or better models?
[7]
Andrews, R., Diederich, J., &#38; Tickle, A. B. (1995). Survey and critique of techniques for extracting rules from trained artificial neural networks. <i>Knowledge-based systems</i>.
[8]
Andrzejak, A., Langner, F., &#38; Zabala, S. (2013). Interpretable models from distributed data via merging of decision trees. In <i>Computational Intelligence and Data Mining (CIDM), 2013 IEEE Symposium on</i>: IEEE.
[9]
Angelov, P. &#38; Soares, E. (2019). Towards explainable deep neural networks (xdnn). <i>arXiv preprint arXiv:1912.02523</i>.
[10]
Askham, N., Cook, D., Doyle, M., Fereday, H., Gibson, M., Landbeck, U., Lee, R., Maynard, C., Palmerand, G., &#38; Schwarzenbach, J. (2013). The six primary dimensions for data quality assessment. <i>DAMA UK Working Group</i>, (pp. 432-435).
[11]
Augasta, M. G. &#38; Kathirvalavakumar, T. (2012). Reverse engineering the neural networks for rule extraction in classification problems. <i>Neural processing letters</i>.
[12]
Baehrens, D., Schroeter, T., Harmeling, S., Kawanabe, M., Hansen, K., &#38; M. Zoeller, K.-R. (2010). How to explain individual classification decisions. <i>Journal of Machine Learning Research</i>.
[13]
Bahdanau, D., Cho, K., &#38; y. Bengio (2014). Neural machine translation by jointly learning to align and translate. <i>arXiv preprint arXiv:1409.0473</i>.
[14]
Balestriero, R. (2017). Neural decision trees. <i>arXiv preprint arXiv:1702.07360</i>.
[15]
Barakat, N. &#38; Diederich, J. (2004). Learning-based rule-extraction from support vector machines. In <i>The 14th International Conference on Computer Theory and applications ICCTA'2004</i>: not found.
[16]
Barakat, N. H. &#38; Bradley, A. P. (2007). Rule extraction from support vector machines: A sequential covering approach. <i>IEEE Transactions on Knowledge and Data Engineering</i>.
[17]
Barbella, D., Benzaid, S., Christensen, J. M., Jackson, B., Qin, X. V., &#38; Musicant, D. (2009). Understanding support vector machine classifications via a recommender systemlike approach. In <i>DMIN</i>.
[18]
Bastani, O., Kim, C., &#38; Bastani, H. (2017). Interpreting blackbox models via model extraction. <i>arXiv preprint arXiv:1705.08504</i>.
[19]
Bengio, Y. &#38; Pearson, J. (2016). When ai goes wrong we won't be able to ask it why.
[20]
Berkson, J. (1953). A statistically precise and relatively simple method of estimating the bio-assay with quantal response, based on the logistic function. <i>Journal of the American Statistical Association</i>.
[21]
Bertsimas, D., Chang, A., &#38; Rudin, C. (2011). Ordered rules for classification: A discrete optimization approach to associative classification. In <i>SUBMITTED TO THE ANNALS OF STATISTICS</i>: Citeseer.
[22]
Bhatt, U., Ravikumar, P., &#38; J. M. F. Moura, J. (2019). Towards aggregating weighted feature attributions. <i>arXiv preprint arXiv:1901.10040</i>.
[23]
Bien, J. &#38; Tibshirani, R. (2009). Classification by set cover: The prototype vector machine. <i>arXiv preprint arXiv:0908.2284</i>.
[24]
Bien, J. &#38; Tibshirani, R. (2011). Prototype selection for interpretable classification. <i>The Annals of Applied Statistics</i>.
[25]
Biran, O. &#38; Cotton, C. (2017). Explanation and Justification in Machine Learning: A survey. In <i>IJCAI-17 Workshop on Explainable AI (XAI)</i>.
[26]
Biran, O. &#38; McKeown, K. R. (2017). Human-centric justification of machine learning predictions. In <i>IJCAI</i>.
[27]
Biswas, S. K., Chakraborty, M., Purkayastha, B., Roy, P., &#38; Thounaojam, D. M. (2017). Rule extraction from training data using neural network. <i>International Journal on Artificial Intelligence Tools</i>.
[28]
Bohanec, M., Borvstnar, M. K., &#38; Robnik-vSikonja, M. (2017). Explaining machine learning models in sales predictions. <i>Expert Systems with Applications</i>.
[29]
Bojars, U. &#38; Breslin, J. G. (2020). Semantically-interlinked online communities.
[30]
Boz, O. (2002). Extracting decision trees from trained neural networks. In <i>Proceedings of the eighth ACM SIGKDD international conference on Knowledge discovery and data mining</i>: ACM.
[31]
Breiman, L. (2017). <i>Classification and regression trees</i>. Routledge.
[32]
Brickley, D. &#38; Miller, L. (2020). The foaf project.
[33]
Burkart, N., Huber, M. F., &#38; Faller, P. (2019). Forcing interpretability for deep neural networks through rule-based regularization. In <i>2019 18th IEEE International Conference On Machine Learning And Applications (ICMLA)</i> (pp. 700-705).: IEEE.
[34]
Byrum, J. (2017). The challenges for artificial intelligence in agriculture.
[35]
Cambridge (2020). <i>The Cambridge dictionary of psychology.</i> Cambridge University Press.
[36]
Caruana, R., Lou, Y., Gehrke, J., Koch, P., Sturm, M., &#38; Elhadad, N. (2015). Intelligible models for healthcare: Predicting pneumonia risk and hospital 30-day readmission. In <i>Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</i>: ACM.
[37]
Charniak, E. (1991). Bayesian networks without tears. <i>AI magazine</i>.
[38]
Chen, D., Fraiberger, S. P., Moakler, R., &#38; Provost, F. (2015). Enhancing transparency and control when drawing data-driven inferences about individuals. <i>Proceedings of 2016 ICML Workshop on Human Interpretability in Machine Learning</i>.
[39]
Chen, J., L&#233;cu&#233;, F., Pan, J. Z., Horrocks, I., &#38; Chen, H. (2018). Knowledge-based transfer learning explanation. <i>CoRR</i>, abs/1807.08372.
[40]
Chen, Y., Ouyang, L., Bao, S., Li, Q., Han, L., Zhang, H., Zhu, B., Xu, M., Liu, J., Ge, Y., et al. (2020). An interpretable machine learning framework for accurate severe vs non-severe covid-19 clinical type classification. <i>medRxiv</i>.
[41]
Clark, P. &#38; Niblett, T. (1989). The cn2 induction algorithm. <i>Machine learning</i>.
[42]
Cleland, S. (2011). Google's 'infringenovation' secrets.
[43]
Cohen, W. (1995). Fast effective rule induction. In <i>Machine Learning Proceedings 1995</i>. Elsevier.
[44]
Confalonieri, R., del Prado, F. M., Agramunt, S., Malagarriga, D., Faggion, D., Weyde, T., &#38; Besold, T. R. (2019). An ontology-based approach to explaining artificial neural networks. <i>CoRR</i>, abs/1906.08362.
[45]
Cortez, P. &#38; Embrechts, M. J. (2011). Opening black box data mining models using sensitivity analysis. In <i>Computational Intelligence and Data Mining (CIDM), 2011 IEEE Symposium on</i>: IEEE.
[46]
Craven, M. &#38; Shavlik, J. W. (1996). Extracting tree-structured representations of trained networks. In <i>Advances in neural information processing systems</i>.
[47]
Cui, Z., Chen, W., He, Y., &#38; Chen, Y. (2015). Optimal action extraction for random forests and boosted trees. In <i>Proceedings of the 21th ACM SIGKDD international conference on knowledge discovery and data mining</i>.
[48]
Datta, A., Sen, S., &#38; Zick, Y. (2016). Algorithmic transparency via quantitative input influence: Theory and experiments with learning systems. In <i>Security and Privacy (SP), 2016 IEEE Symposium on</i>: IEEE.
[49]
Doan, A., Madhavan, J., Domingos, P., &#38; Halevy, A. (2004). <i>Ontology Matching: A Machine Learning Approach</i>, (pp. 385-403). Springer Berlin Heidelberg: Berlin, Heidelberg.
[50]
Doran, D., Schulz, S., &#38; Besold, T. R. (2017). What does explainable ai really mean? a new conceptualization of perspectives. <i>arXiv preprint arXiv:1710.00794</i>.
[51]
Doshi-Velez, F. &#38; Kim, B. (2017). Towards a rigorous science of interpretable machine learning. <i>arXiv preprint arXiv:1702.08608</i>.
[52]
Dosilovi&#263;, F. K., Brci&#263;, M., &#38; Hlupi&#263;, N. (2018). Explainable artificial intelligence: A survey. In <i>2018 41st International convention on information and communication technology, electronics and microelectronics (MIPRO)</i>.
[53]
Dua, D. &#38; Graff, C. (2017). UCI machine learning repository.
[54]
Efron, B., Hastie, T., Johnstone, I., Tibshirani, R., et al. (2004). Least angle regression. <i>The Annals of statistics</i>.
[55]
El-Bekri, N., Kling, J., &#38; Huber, M. F. (2019). A study on trust in black box models and post-hoc explanations. In <i>International Workshop on Soft Computing Models in Industrial and Environmental Applications</i>: Springer.
[56]
Etchells, T. A. &#38; Lisboa, P. J. G. (2006). Orthogonal search-based rule extraction (osre) for trained neural networks: a practical and efficient approach. <i>IEEE transactions on neural networks</i>.
[57]
Europa.eu (2017). Official journal of the european union: Regulations.
[58]
F. Bao, a. Y. H., Liu, J., Chen, Y., Li, Q., Zhang, C., Han, L., Zhu, B., Ge, Y., Chen, S., et al. (2020). Triaging moderate covid-19 and other viral pneumonias from routine blood tests. <i>arXiv preprint arXiv:2005.06546</i>.
[59]
Fan, X., Liu, S., Chen, J., &#38; Henderson, T. C. (2020). An investigation of covid-19 spreading factors with explainable ai techniques. <i>arXiv preprint arXiv:2005.06612</i>.
[60]
Fischer, G., Mastaglio, T., Reeves, B., &#38; Rieman, J. (1990). Minimalist explanations in knowledge-based systems. In <i>Twenty-Third Annual Hawaii International Conference on System Sciences</i>, volume 3 (pp. 309-317 vol.3).
[61]
Fisher, A., Rudin, C., &#38; Dominici, F. (2018). Model class reliance: Variable importance measures for any machine learning model class, from the" rashomon" perspective. <i>arXiv preprint arXiv:1801.01489</i>.
[62]
Freitas, A. (2014). Comprehensible classification models: a position paper. <i>ACM SIGKDD explorations newsletter</i>.
[63]
Friedman, J. H., Popescu, B. E., et al. (2008). Predictive learning via rule ensembles. <i>The Annals of Applied Statistics</i>.
[64]
Friedman, N., Geiger, D., &#38; Goldszmidt, M. (1997). Bayesian network classifiers. <i>Machine learning</i>.
[65]
Fu, L. (1994). Rule generation from neural networks. <i>IEEE Transactions on Systems, Man, and Cybernetics</i>.
[66]
Fung, G., Sandilya, S., &#38; Rao, R. B. (2008). Rule extraction from linear support vector machines via mathematical programming. In <i>Rule Extraction from Support Vector Machines</i>. Springer.
[67]
Geng, Y., Chen, J., Jimenez-Ruiz, E., &#38; Chen, H. (2019). Human-centric transfer learning explanation via knowledge graph [extended abstract].
[68]
Gilpin, L. H., Bau, D., Yuan, B. Z., Bajwa, A., Specter, M., &#38; Kagal, L. (2018). Explaining explanations: An overview of interpretability of machine learning. In <i>2018 IEEE 5th International Conference on data science and advanced analytics (DSAA)</i>.
[69]
Gkatzia, D., Lemon, O., &#38; Rieser, V. (2016). Natural language generation enhances human decision-making with uncertain information. <i>arXiv preprint arXiv:1606.03254</i>.
[70]
Goldstein, A., Kapelner, A., Bleich, J., &#38; Pitkin, E. (2015). Peeking inside the black box: Visualizing statistical learning with plots of individual conditional expectation. <i>Journal of Computational and Graphical Statistics</i>.
[71]
Goodman, B. &#38; Flaxman, S. (2016). Eu regulations on algorithmic decision-making and a "right to explanation". In <i>ICML workshop on human interpretability in machine learning (WHI 2016), New York, NY</i>.
[72]
Gruber, T. R. et al. (1993). A translation approach to portable ontology specifications. <i>Knowledge acquisition</i>, 5(2), 199-221.
[73]
Gudivada, V., Apon, A., &#38; Ding, J. (2017). Data quality considerations for big data and machine learning: Going beyond data cleaning and transformations. <i>International Journal on Advances in Software</i>, 10(1), 1-20.
[74]
Guidotti, R., Monreale, A., Ruggieri, S., Pedreschi, D., Turini, F., &#38; Giannotti, F. (2018a). Local rule-based explanations of black box decision systems. <i>arXiv preprint arXiv:1805.10820</i>.
[75]
Guidotti, R., Monreale, A., Ruggieri, S., Turini, F., Giannotti, F., &#38; Pedreschi, D. (2018b). A survey of methods for explaining black box models. <i>ACM Comput. Surv</i>.
[76]
Gunning, D. (2017). Explainable artificial intelligence (xai). <i>Defense Advanced Research Projects Agency (DARPA)</i>.
[77]
Gurumoorthy, K. S., Dhurandhar, A., &#38; Cecchi, G. (2017). Protodash: Fast interpretable prototype selection. <i>arXiv preprint arXiv:1707.01212</i>.
[78]
Hall, P., Gill, N., Kurka, M., &#38; Phan, W. (2017a). Machine learning interpretability with h2o driverless ai. <i>H2O.ai</i>.
[79]
Hall, P., Phan, W., &#38; Ambati, S. (2017b). Ideas on interpreting machine learning.
[80]
Hara, S. &#38; Hayashi, K. (2016). Making tree ensembles interpretable. <i>arXiv preprint arXiv:1606.05390</i>.
[81]
Hayashi, Y. (2013). Neural network rule extraction by a new ensemble concept and its theoretical and historical background: A review. <i>International Journal of Computational Intelligence and Applications</i>.
[82]
Helfert, M. &#38; Ge, M. (2016). Big data quality-towards an explanation model in a smart city context. In <i>proceedings of 21st International Conference on Information Quality, Ciudad Real, Spain</i>.
[83]
Hendricks, L. A., Akata, Z., Rohrbach, M., Donahue, J., Schiele, B., &#38; Darrell, T. (2016). Generating visual explanations. In <i>European Conference on Computer Vision</i>: Springer.
[84]
Henelius, A., Puolam&#228;ki, K., Bostr&#246;m, H., Asker, L., &#38; Papapetrou, P. (2014). A peek into the black box: exploring classifiers by randomization. <i>Data mining and knowledge discovery</i>.
[85]
Henelius, A., Puolam&#228;ki, K., &#38; Ukkonen, A. (2017). Interpreting classifiers through attribute interactions in datasets. In <i>2017 ICML Workshop on Human Interpretability in Machine Learning (WHI)</i>.
[86]
Hepp, M. (2020). Good relations.
[87]
Herman, B. (2017). The promise and peril of human evaluation for model interpretability. <i>arXiv preprint arXiv:1711.07414</i>.
[88]
Hilton, D. J. (1990). Conversational processes and causal explanation. <i>Psychological Bulletin</i>.
[89]
Hinton, G. &#38; Frosst, N. (2017). Distilling a neural network into a soft decision tree. In <i>Comprehensibility and Explanation in AI and ML (CEX), AI*IA</i>.
[90]
Hoehndorf, R. (2010). What is an upper level ontology? <i>Ontogenesis</i>.
[91]
Hoffman, R., Mueller, S., Klein, G., &#38; Litman, J. (2018). Metrics for explainable ai: Challenges and prospects. <i>arXiv preprint arXiv:1812.04608</i>.
[92]
Holte, R. C. (1993). Very simple classification rules perform well on most commonly used datasets. <i>Machine learning</i>.
[93]
Holzinger, A., Kickmeier-Rust, M., &#38; M&#252;ller, H. (2019a). Kandinsky patterns as iq-test for machine learning. In <i>International Cross-Domain Conference for Machine Learning and Knowledge Extraction</i> (pp. 1-14).: Springer.
[94]
Holzinger, A., Langs, G., Denk, H., Zatloukal, K., &#38; M&#252;ller, H. (2019b). Causability and explainabilty of artificial intelligence in medicine. <i>Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery</i>.
[95]
Holzinger, A., Plass, M., Holzinger, K., Crisan, G. C., Pintea, C. M., &#38; Palade, V. (2017). A glass-box interactive machine learning approach for solving np-hard problems with the human-in-the-loop. <i>arXiv preprint arXiv:1708.01104</i>.
[96]
Holzinger, A., Plass, M., Kickmeier-Rust, M., Holzinger, K., Cri&#351;an, G. C., Pintea, C. M., &#38; Palade, V. (2019c). Interactive machine learning: experimental evidence for the human in the algorithmic loop. <i>Applied Intelligence</i>, 49(7), 2401-2414.
[97]
Hoyt, R. E., Snider, D., Thompson, C., &#38; Mantravadi, S. (2016). Ibm watson analytics: Automating visualization, descriptive, and predictive statistics. <i>JMIR Public Health Surveill</i>, 2(2), e157.
[98]
Huysmans, J., Baesens, B., &#38; Vanthienen, J. (2006). Iter: an algorithm for predictive regression rule extraction. In <i>International Conference on Data Warehousing and Knowledge Discovery</i>: Springer.
[99]
Huysmans, J., Dejaeger, K., Mues, C., Vanthienen, J., &#38; Baesens, B. (2011). An empirical evaluation of the comprehensibility of decision table, tree and rule based predictive models. <i>Decision Support Systems</i>.
[100]
Jain, S. &#38; Wallace, B. C. (2019). Attention is not explanation. <i>arXiv preprint arXiv:1902.10186</i>.
[101]
Jiang, T. &#38; Owen, A. B. (2002). Quasi-regression for visualization and interpretation of black box functions.
[102]
Johansson, U., K&#246;nig, R., &#38; Niklasson, L. (2004). The truth is in there-rule extraction from opaque models using genetic programming. In <i>FLAIRS Conference</i>: Miami Beach, FL.
[103]
Kabra, M., Robie, A., &#38; Branson, K. (2015). Understanding classifier errors by examining inuential neighbors. In <i>Proceedings of the IEEE conference on computer vision and pattern recognition</i>.
[104]
Kaggle (2017). The state of data science and machine learning.
[105]
Kamruzzaman, S. (2010). Rex: An efficient rule generator. <i>arXiv preprint arXiv:1009.4988</i>.
[106]
Kass, R. &#38; Finin, T. (1988). The Need for User Models in Generating Expert System Explanations. <i>International Journal of Expert Systems</i>, 1(4).
[107]
Kim, B., Khanna, R., &#38; Koyejo, O. O. (2016). Examples are not enough, learn to criticize! criticism for interpretability. In <i>Advances in Neural Information Processing Systems</i>.
[108]
Kim, B., Rudin, C., &#38; Shah, J. A. (2014). The bayesian case model: A generative approach for case-based reasoning and prototype classification. In <i>Advances in Neural Information Processing Systems</i>.
[109]
Kim, B., Shah, J. A., &#38; Doshi-Velez, F. (2015). Mind the gap: A generative approach to interpretable feature selection and extraction. In <i>Advances in Neural Information Processing Systems</i>.
[110]
Kittler, J. (1986). Feature selection and extraction. <i>Handbook of Pattern Recognition and Image Processing</i>.
[111]
Koh, P. W. &#38; Liang, P. (2017). Understanding black-box predictions via influence functions. <i>arXiv preprint arXiv:1703.04730</i>.
[112]
Kramer, M. A. (1991). Nonlinear principal component analysis using autoassociative neural networks. <i>AIChE journal</i>, 37(2), 233-243.
[113]
Krause, J., Perer, A., &#38; Ng, K. (2016). Interacting with predictions: Visual inspection of black-box machine learning models. In <i>Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems</i>: ACM.
[114]
Lage, I., Chen, E., He, J., Narayanan, M., Kim, B., Gershman, S., &#38; Doshi-Velez, F. (2019). An evaluation of the human-interpretability of explanation. <i>arXiv preprint arXiv:1902.00006</i>.
[115]
Lakkaraju, H., Bach, S. H., &#38; Leskovec, J. (2016). Interpretable decision sets: A joint framework for description and prediction. In <i>Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</i>: ACM.
[116]
Lakkaraju, H., Kamar, E., Caruana, R., &#38; Leskovec, J. (2017). Interpretable &#38; explorable approximations of black box models. <i>arXiv preprint arXiv:1707.01154</i>.
[117]
Lakkaraju, H., Kamar, E., Caruana, R., &#38; Leskovec, J. (2019). Faithful and customizable explanations of black box models. In <i>Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society</i>.
[118]
Lash, M. T., Lin, Q., Street, W. N., &#38; Robinson, J. G. (2017). A budget-constrained inverse classification framework for smooth classifiers. In <i>2017 IEEE International Conference on Data Mining Workshops (ICDMW)</i>.
[119]
Laugel, T., Lesot, M. J., Marsala, C., Renard, X., &#38; Detyniecki, M. (2017). Inverse classification for comparison-based interpretability in machine learning. <i>arXiv preprint arXiv:1712.08443</i>.
[120]
Laugel, T., Renard, X., Lesot, M., Marsala, C., &#38; Detyniecki, M. (2018). Defining locality for surrogates in post-hoc interpretablity. <i>arXiv preprint arXiv:1806.07498</i>.
[121]
L&#233;cu&#233;, F., Abeloos, B., Anctil, J., Bergeron, M., Dalla-Rosa, D., Corbeil-Letourneau, S., Martet, F., Pommellet, T., Salvan, L., Veilleux, S., &#38; Ziaeefard, M. (2019). Thales xai platform: Adaptable explanation of machine learning systems - a knowledge graphs perspective. In <i>ISWC Satellites</i>.
[122]
Lei, J., G'Sell, M., Rinaldo, A., Tibshirani, R. J., &#38; Wasserman, L. (2018). Distribution-free predictive inference for regression. <i>Journal of the American Statistical Association</i>.
[123]
Lei, T., Barzilay, R., &#38; Jaakkola, T. (2016). Rationalizing neural predictions. <i>arXiv preprint arXiv:1606.04155</i>.
[124]
Lent, M. V., Fisher, W., &#38; Mancuso, M. (2004). An explainable artificial intelligence system for small-unit tactical behavior. In <i>Proceedings of the national conference on artificial intelligence</i>.
[125]
Letham, B., Rudin, C., McCormick, T. H., &#38; Madigan, D. (2012). Building interpretable classifiers with rules using bayesian analysis. <i>Department of Statistics Technical Report tr609, University of Washington</i>.
[126]
Letham, B., Rudin, C., McCormick, T. H., &#38; Madigan, D. (2015). Interpretable classifiers using rules and bayesian analysis: Building a better stroke prediction model. <i>The Annals of Applied Statistics</i>.
[127]
Lipton, Z., Kale, D., &#38; Wetzel, R. (2016). Modeling missing data in clinical time series with rnns. <i>arXiv preprint arXiv:1606.04130</i>.
[128]
Lipton, Z. C. (2017). The doctor just won't accept that! <i>arXiv preprint arXiv:1711.08037</i>.
[129]
Lipton, Z. C. (2018). The mythos of model interpretability. <i>Queue</i>, 16(3), 31-57.
[130]
Looveren, A. V. &#38; Klaise, J. (2019). Interpretable counterfactual explanations guided by prototypes. <i>arXiv preprint arXiv:1907.02584</i>.
[131]
Lou, Y., Caruana, R., &#38; Gehrke, J. (2012). Intelligible models for classification and regression. In <i>Proceedings of the 18th ACM SIGKDD international conference on Knowledge discovery and data mining</i>: ACM.
[132]
Lou, Y., Caruana, R., Gehrke, J., &#38; Hooker, G. (2013). Accurate intelligible models with pairwise interactions. In <i>Proceedings of the 19th ACM SIGKDD international conference on Knowledge discovery and data mining</i>: ACM.
[133]
Lu, H., Setiono, R., &#38; Liu, H. (1995). Neurorule: A connectionist approach to data mining. In <i>Proceedings of the 21st VLDB Conference Zurich, Switzerland</i>.
[134]
Lundberg, S. M. &#38; Lee, S. (2017). A unified approach to interpreting model predictions. In I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, &#38; R. Garnett (Eds.), <i>Advances in Neural Information Processing Systems 30</i> (pp. 4765-4774). Curran Associates, Inc.
[135]
Luong, M. T., Pham, H., &#38; Manning, C. D. (2015). Effective approaches to attention-based neural machine translation. <i>arXiv preprint arXiv:1508.04025</i>.
[136]
Maedche, A. &#38; Staab, S. (2001). Ontology learning for the semantic web. <i>IEEE Intelligent Systems</i>, 16, 72-79.
[137]
Mahajan, D., Tan, C., &#38; Sharma, A. (2019). Preserving causal constraints in counterfactual explanations for machine learning classifiers. <i>arXiv preprint arXiv:1912.03277</i>.
[138]
Malioutov, D. M., Varshney, K. R., Emad, A., &#38; Dash, S. (2017). Learning interpretable classification rules with boolean compressed sensing. In <i>Transparent Data Mining for Big and Small Data</i>. Springer.
[139]
Markowska-Kaczmar, U. &#38; Chumieja, M. (2004). Discovering the mysteries of neural networks. <i>International Journal of Hybrid Intelligent Systems</i>.
[140]
Martens, D., Backer, M. D., Haesen, R., Vanthienen, J., Snoeck, M., &#38; Baesens, B. (2007a). Classification with ant colony optimization. <i>IEEE Transactions on Evolutionary Computation</i>.
[141]
Martens, D., Baesens, B., &#38; Gestel, T. V. (2009). Decompositional rule extraction from support vector machines by active learning. <i>IEEE Transactions on Knowledge and Data Engineering</i>.
[142]
Martens, D., Baesens, B., Gestel, T. V., &#38; Vanthienen, J. (2007b). Comprehensible credit scoring models using rule extraction from support vector machines. <i>European journal of operational research</i>.
[143]
Martens, D., Huysmans, J., Setiono, R., Vanthienen, J., &#38; Baesens, B. (2008). Rule extraction from support vector machines: an overview of issues and application in credit scoring. <i>Rule extraction from support vector machines</i>.
[144]
Martens, D. &#38; Provost, F. (2014). Explaining data-driven document classifications. <i>Mis Quarterly</i>.
[145]
Martens, D., Vanthienen, J., Verbeke, W., &#38; Baesens, B. (2011). Performance of classification models from a user perspective. <i>Decision Support Systems</i>.
[146]
Mashayekhi, M. &#38; Gras, R. (2017). Rule extraction from decision trees ensembles: New algorithms based on heuristic search and sparse group lasso methods. <i>International Journal of Information Technology &#38; Decision Making</i>.
[147]
Massachusetts Institute of Technology (2017). The moral machine.
[148]
McGuinness, D. L., Ding, L., da Silva, P., &#38; Chang, C. (2007). Pml 2: A modular explanation interlingua. In <i>ExaCt</i>.
[149]
Mead, A. (1992). Review of the development of multidimensional scaling methods. <i>Journal of the Royal Statistical Society: Series D (The Statistician)</i>, 41(1), 27-39.
[150]
Meinshausen, N. (2010). Node harvest. <i>The Annals of Applied Statistics</i>.
[151]
Melis, D. A. &#38; Jaakkola, T. (2018). Towards robust interpretability with self-explaining neural networks. In <i>Advances in Neural Information Processing Systems</i>.
[152]
Miller, T. (2019). Explanation in artificial intelligence: Insights from the social sciences. <i>Artificial Intelligence</i>, 267, 1-38.
[153]
Mohammed, O., Benlamri, R., &#38; Fong, S. (2012). Building a diseases symptoms ontology for medical diagnosis: An integrative approach. In <i>The First International Conference on Future Generation Communication Technologies</i>.
[154]
Molnar, C. (2018). A guide for making black box models explainable. URL: https://christophm.github.io/interpretable-ml-book.
[155]
Montavon, G., Lapuschkin, S., Binder, A., Samek, W., &#38; M&#252;ller, K. R. (2017). Explaining nonlinear classification decisions with deep taylor decomposition. <i>Pattern Recognition</i>.
[156]
Montavon, G., Samek, W., &#38; M&#252;ller, K. R. (2018). Methods for interpreting and understanding deep neural networks. <i>Digital Signal Processing</i>.
[157]
Murdoch, J., Singh, C., Kumbier, K., Abbasi-Asl, R., &#38; Yu, B. (2019). Interpretable machine learning: definitions, methods, and applications. <i>arXiv preprint arXiv:1901.04592</i>.
[158]
Navigli, R. &#38; Velardi, P. (2004). Learning domain ontologies from document warehouses and dedicated web sites. <i>Computational Linguistics</i>, 30(2), 151-179.
[159]
Ninama, H. (2013). Ensemble approach for rule extraction in data mining. <i>Golden Reaserch Thoughts</i>.
[160]
Odajima, K., Hayashi, Y., Tianxia, G., &#38; Setiono, R. (2008). Greedy rule generation from discrete data and its use in neural network rule extraction. <i>Neural Networks</i>.
[161]
Otero, F. E. B. &#38; Freitas, A. (2016). Improving the interpretability of classification rules discovered by an ant colony algorithm: Extended results. <i>Evolutionary Computation</i>.
[162]
Panigutti, C., Perotti, A., &#38; Pedreschi, D. (2020). Doctor xai: An ontology-based approach to black-box sequential data classification explanations. In <i>Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency</i>, FAT* '20 (pp. 629-639). New York, NY, USA: Association for Computing Machinery.
[163]
Park, D. H., Hendricks, L. A., Akata, Z., Schiele, B., Darrell, T., &#38; Rohrbach, M. (2016). Attentive explanations: Justifying decisions and pointing to the evidence. <i>arXiv preprint arXiv:1612.04757</i>.
[164]
Phillips, R. L., Chang, K. H., &#38; Friedler, S. A. (2017). Interpretable active learning. <i>arXiv preprint arXiv:1708.00049</i>.
[165]
Plumb, G., Molitor, D., &#38; Talwalkar, A. S. (2018). Model agnostic supervised local explanations. In <i>Advances in Neural Information Processing Systems</i>.
[166]
Poursabzi-Sangdeh, F., Goldstein, D. G., Hofman, J. M., Vaughan, J. W., &#38; Wallach, H. (2018). Manipulating and measuring model interpretability. <i>arXiv preprint arXiv:1802.07810</i>.
[167]
Poyiadzi, R., Sokol, K., Santos-Rodriguez, R., De Bie, T., &#38; Flach, P. (2020). Face: feasible and actionable counterfactual explanations. In <i>Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society</i> (pp. 344-350).
[168]
Publio, G. C., Esteves, D., Lawrynowicz, A., ce Panov, P., Soldatova, L., Soru, T., Vanschoren, J., &#38; Zafar, H. (2018). Ml-schema: Exposing the semantics of machine learning with schemas and ontologies.
[169]
Quinlan, J. R. (1986). Induction of decision trees. <i>Machine learning</i>.
[170]
Quinlan, J. R. (1996). Bagging, boosting, and c4.5. In <i>AAAI/IAAI, Vol. 1</i>.
[171]
Quinlan, J. R. (2014). <i>C4.5: programs for machine learning</i>. Elsevier.
[172]
Raimond, Y., Abdallah, S., Sandler, M., &#38; Giasson, F. (2007). The music ontology. In <i>Proceedings of the 8th International Conference on Music Information Retrieval (ISMIR)</i>.
[173]
Raimond, Y., Abdallah, S., Sandler, M., &#38; Giasson, F. (2020). The music ontology.
[174]
Rezaul, K., D&#195;Phmen, T., Rebholz-Schuhmann, D., Decker, S., Cochez, M., &#38; Beyan, O. (2020). Deepcovidexplainer: Explainable covid-19 predictions based on chest x-ray images. <i>arXiv</i>, (pp. arXiv-2004).
[175]
Ribeiro, M. T., Singh, S., &#38; Guestrin, C. (2016a). Model-agnostic interpretability of machine learning. <i>arXiv preprint arXiv:1606.05386</i>.
[176]
Ribeiro, M. T., Singh, S., &#38; Guestrin, C. (2016b). Why should i trust you?: Explaining the predictions of any classifier. In <i>Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</i>: ACM.
[177]
Ribeiro, M. T., Singh, S., &#38; Guestrin, C. (2016c). Why should i trust you?: Explaining the predictions of any classifier. In <i>Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining</i>: ACM.
[178]
Ribeiro, M. T., Singh, S., &#38; Guestrin, C. (2018). Anchors: High-precision model-agnostic explanations. In <i>Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence (AAAI)</i>.
[179]
Robnik, M. &#38; Kononenko, I. (2008). Explaining classifications for individual instances. <i>IEEE Transactions on Knowledge and Data Engineering</i>.
[180]
Robnik-vSikonja, M. &#38; Kononenko, I. (2008). Explaining classifications for individual instances. <i>IEEE Transactions on Knowledge and Data Engineering</i>.
[181]
Rudin, C. (2018). Please stop explaining black box models for high stakes decisions. <i>CoRR</i>.
[182]
R&#252;ping, S. (2005). Learning with local models. In <i>Local Pattern Detection</i> (pp. 153-170).
[183]
R&#252;ping, S. (2006). Learning interpretable models. <i>Doctoral Dissertation, University of Dortmund</i>.
[184]
Rush, A. M., Chopra, S., &#38; Weston, J. (2015). A neural attention model for abstractive sentence summarization. <i>arXiv preprint arXiv:1509.00685</i>.
[185]
Russell, C. (2019). Efficient search for diverse coherent explanations. In <i>Proceedings of the Conference on Fairness, Accountability, and Transparency</i> (pp. 20-28).
[186]
Saabas, A. (2015). Treeinterpreter. https://github.com/andosa/treeinterpreter.
[187]
Samek, W., Montavon, G., Vedaldi, A., Hansen, L. K., &#38; M&#252;ller, K., Eds. (2019). <i>Explainable AI: Interpreting, Explaining and Visualizing Deep Learning</i>. Springer.
[188]
Samek, W., Wiegand, T., &#38; M&#252;ller, K. R. (2017). Explainable artificial intelligence: Understanding, visualizing and interpreting deep learning models. <i>arXiv preprint arXiv:1708.08296</i>.
[189]
Sarker, M. K., Xie, N., Doran, D., Raymer, M., &#38; Hitzler, P. (2017). Explaining trained neural networks with semantic web technologies: First steps.
[190]
Schaaf, N. &#38; Huber, M. F. (2019). Enhancing decision tree based interpretation of deep neural networks through l1-orthogonal regularization. <i>arXiv preprint arXiv:1904.05394</i>.
[191]
Schetinin, V., Fieldsend, J. E., Partridge, D., Coats, T. J., Krzanowski, W. J., Everson, R. M., Bailey, T. C., &#38; Hernandez, A. (2007). Confident interpretation of bayesian decision tree ensembles for clinical applications. <i>IEEE Transactions on Information Technology in Biomedicine</i>.
[192]
Schmidt, P. &#38; Biessmann, F. (2019). Quantifying interpretability and trust in machine learning systems. <i>arXiv preprint arXiv:1901.08558</i>.
[193]
Schmitz, G., Aldrich, C., &#38; Gouws, F. S. (1999). Ann-dt: an algorithm for extraction of decision trees from artificial neural networks. <i>IEEE Transactions on Neural Networks</i>.
[194]
Selvaraju, R. R., Das, A., Vedantam, R., Cogswell, M., Parikh, D., &#38; Batra, D. (2016). Grad-cam: Why did you say that? visual explanations from deep networks via gradient-based localization. <i>arXiv preprint arXiv:1610.02391</i>.
[195]
Sestito, S. &#38; Dillon, T. (1992). Automated knowledge acquisition of rules with continuously valued attributes. In <i>Proceedings of the 12th international conference on expert systems and their applications, 1992</i>.
[196]
Sethi, K. K., Mishra, D. K., &#38; Mishra, B. (2012). Extended taxonomy of rule extraction techniques and assessment of kdruleex. <i>International Journal of Computer Applications</i>.
[197]
Setiono, R., Azcarraga, A., &#38; Hayashi, Y. (2014). Mofn rule extraction from neural networks trained with augmented discretized input. In <i>Neural Networks (IJCNN), 2014 International Joint Conference on</i>: IEEE.
[198]
Setiono, R., Baesens, B., &#38; Mues, C. (2008). Recursive neural network rule extraction for data with mixed attributes. <i>IEEE Transactions on Neural Networks</i>.
[199]
Setiono, R. &#38; Liu, H. (1997). Neurolinear: From neural networks to oblique decision rules. <i>Neurocomputing</i>.
[200]
Shapley, L. S. (1951). <i>Notes on the n-Person Game-II: The Value of an n-Person Game</i>. Technical report, U.S. Air Force, Project Rand.
[201]
Shrikumar, A., Greenside, P., Shcherbina, A., &#38; Kundaje, A. (2016). Not just a black box: Learning important features through propagating activation differences. In <i>33rd International Conference on Machine Learning</i>.
[202]
Si, Z. &#38; Zhu, S. C. (2013). Learning and-or templates for object recognition and detection. <i>IEEE transactions on pattern analysis and machine intelligence</i>.
[203]
Smilkov, D., Thorat, N., Kim, B., Vi&#233;gas, F., &#38; Wattenberg, M. (2017). Smoothgrad: removing noise by adding noise. <i>arXiv preprint arXiv:1706.03825</i>.
[204]
Strumbelj, E., Bosni&#263;, Z., Kononenko, I., Zakotnik, B., &#38; Kuhar, C. (2010). Explanation and reliability of prediction models: the case of breast cancer recurrence. <i>Knowledge and information systems</i>.
[205]
Strumbelj, E. &#38; Kononenko, I. (2014). Explaining prediction models and individual predictions with feature contributions. <i>Knowledge and information systems</i>.
[206]
Su, G., Wei, D., Varshney, K. R., &#38; Malioutov, D. M. (2015). Interpretable two-level boolean rule learning for classification. <i>arXiv preprint arXiv:1511.07361</i>.
[207]
Su, G., Wei, D., Varshney, K. R., &#38; Malioutov, D. M. (2016). Learning sparse two-level boolean rules. In <i>2016 IEEE 26th International Workshop on Machine Learning for Signal Processing (MLSP)</i>: IEEE.
[208]
Subianto, M. &#38; Siebes, A. (2007). Understanding discrete classifiers with a case study in gene prediction. In <i>Seventh IEEE International Conference on Data Mining 2007</i> (pp. 661-666).: IEEE.
[209]
Sundararajan, M., Taly, A., &#38; Yan, Q. (2016). Gradients of counterfactuals. <i>arXiv preprint arXiv:1611.02639</i>.
[210]
Swartout, W., Paris, C., &#38; Moore, J. (1991). Explanations in knowledge systems: design for explainable expert systems. <i>IEEE Expert</i>, 6(3), 58-64.
[211]
Taha, I. &#38; Ghosh, J. (1996). Three techniques for extracting rules from feedforward networks. <i>Intelligent Engineering Systems Through Artificial Neural Networks</i>.
[212]
Tibshirani, R. (1996). Regression shrinkage and selection via the lasso. <i>Journal of the Royal Statistical Society: Series B (Methodological)</i>.
[213]
Tjoa, E. &#38; Guan, C. (2019). A survey on explainable artificial intelligence (xai): Towards medical xai. <i>arXiv preprint arXiv:1907.07374</i>.
[214]
Tolomei, G., Silvestri, F., Haines, A., &#38; Lalmas, M. (2017). Interpretable predictions of tree-based ensembles via actionable feature tweaking. In <i>Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</i>: ACM.
[215]
Tsymbal, A., Zillner, S., &#38; Huber, M. (2007). Ontology - Supported Machine Learning and Decision Support in Biomedicine. In <i>International Conference on Data Integration in the Life Sciences</i>, volume 4544 (pp. 156-171).
[216]
Turner, R. (2016). A model explanation system. In <i>2016 IEEE 26th International Workshop on Machine Learning for Signal Processing (MLSP)</i>.
[217]
Tversky, A. &#38; Kahneman, D. (1974). Judgment under uncertainty: Heuristics and biases. <i>Science</i>.
[218]
Tversky, A. &#38; Kahneman, D. (1981). The framing of decisions and the psychology of choice. <i>Science</i>.
[219]
Ustun, B. &#38; Rudin, C. (2014). Methods and models for interpretable linear classification. <i>arXiv preprint arXiv:1405.4047</i>.
[220]
Ustun, B. &#38; Rudin, C. (2016). Supersparse linear integer models for optimized medical scoring systems. <i>Machine Learning</i>.
[221]
Ustun, B. &#38; Rudin, C. (2017). Optimized risk scores. In <i>Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</i>: ACM.
[222]
van der Maaten, L. &#38; Hinton, G. (2008). Visualizing data using t-sne. <i>Journal of machine learning research</i>, 9(Nov), 2579-2605.
[223]
Vedantam, R., Bengio, S., Murphy, K., Parikh, D., &#38; Chechik, G. (2017). Context-aware captions from context-agnostic supervision. In <i>Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition</i>.
[224]
Verbeke, W., Martens, D., Mues, C., &#38; Baesens, B. (2011). Building comprehensible customer churn prediction models with advanced rule induction techniques. <i>Expert Systems with Applications</i>.
[225]
Voosen, P. (2017). How AI detectives are cracking open the black box of deep learning. <i>Science Magazine</i>.
[226]
W3C (2012a). <i>Good Ontologies</i>. W3C recommendation, W3C. https://www.w3.org/wiki/Good Ontologies.
[227]
W3C (2012b). <i>OWL 2 Web Ontology Language Document Overview (Second Edition)</i>. W3C recommendation, W3C. https://www.w3.org/TR/2012/REC-owl2-overview-20121211/.
[228]
W3C (2014). <i>Ressource Description Framework(RDF)</i>. W3C recommendation, W3C. https://www.w3.org/RDF/.
[229]
Wachter, S., Mittelstadt, B., &#38; Russell, C. (2018). Counterfactual explanations without opening the black box: Automated decisions and the gdpr. <i>Harvard Journal of Law &#38; Technology</i>, 31(2).
[230]
Wang, F. &#38; Rudin, C. (2014). Falling rule lists. <i>arXiv preprint arXiv:1411.5899</i>.
[231]
Wang, F. &#38; Rudin, C. (2015). Falling rule lists. In <i>18th International Conference on Artificial Intelligence and Statistics (AISTATS)</i>.
[232]
Wang, J., Fujimaki, R., &#38; Motohashi, Y. (2015a). Trading interpretability for accuracy: Oblique treed sparse additive models. In <i>Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</i>: ACM.
[233]
Wang, R. Y. &#38; Strong, D. M. (1996). Beyond accuracy: What data quality means to data consumers. <i>Journal of management information systems</i>, 12(4), 5-33.
[234]
Wang, T., Rudin, C., Doshi-Velez, F., Liu, Y., Klampfl, E., &#38; MacNeille, P. (2015b). Or's of and's for interpretable classification, with application to context-aware recommender systems. <i>arXiv preprint arXiv:1504.07614</i>.
[235]
Wang, T., Rudin, C., Velez-Doshi, F., Liu, Y., Klamp, E., &#38; MacNeille, P. (2016). Bayesian rule sets for interpretable classification. In <i>Data Mining (ICDM), 2016 IEEE 16th International Conference on</i>: IEEE.
[236]
Weiner, J. (1980). Blah, a system which explains its reasoning. <i>Artificial intelligence</i>, 15(1-2), 19-48.
[237]
Weller, A. (2017). Challenges for transparency. <i>arXiv preprint arXiv:1708.01870</i>.
[238]
West, J., Ventura, D., &#38; Warnick, S. (2007). Spring research presentation: A theoretical foundation for inductive transfer. Retrieved 2007-08-05.
[239]
Wiegreffe, S. &#38; Pinter, Y. (2019). Attention is not not explanation. <i>arXiv preprint arXiv:1908.04626</i>.
[240]
Wold, S., Esbense, K., &#38; Geladi, P. (1987). Principal component analysis. <i>Chemometrics and intelligent laboratory systems</i>, 2(1-3), 37-52.
[241]
Wong, W., Liu, W., &#38; Bennamoun, M. (2011). Ontology learning from text: A look back and into the future. <i>ACM Computing Surveys - CSUR</i>, 44, 1-36.
[242]
Wu, M., Hughes, M. C., Parbhoo, S., Zazzi, M., Roth, V., &#38; Doshi-Velez, F. (2018). Beyond sparsity: Tree regularization of deep models for interpretability. In <i>Thirty-Second AAAI Conference on Artificial Intelligence</i>.
[243]
Xu, K., Ba, J., Kiros, R., Cho, K., Courville, A., Salakhudinov, R., Zemel, R., &#38; Bengio, Y. (2015a). Show, attend and tell: Neural image caption generation with visual attention. In <i>International conference on machine learning</i>.
[244]
Xu, N., Jiangping, W., Qi, G., Huang, T., &#38; Lin, W. (2015b). Ontological random forests for image classification. <i>International Journal of Information Retrieval Research</i>, 5, 61-74.
[245]
Yang, C., Rangarajan, A., &#38; Ranka, S. (2018a). Global model interpretation via recursive partitioning. <i>arXiv preprint arXiv:1802.04253</i>.
[246]
Yang, H., Rudin, C., &#38; Seltzer, M. (2016). Scalable bayesian rule lists. unpublished.
[247]
Yang, Y., Morillo, I. G., &#38; Hospedales, T. M. (2018b). Deep neural decision trees. <i>arXiv preprint arXiv:1806.06988</i>.
[248]
Yin, M., Vaughan, J. W., &#38; Wallach, H. (2019). Understanding the effect of accuracy on trust in machine learning models. In <i>Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems</i>: ACM.
[249]
Yin, X. &#38; Han, J. (2003). Cpar: Classification based on predictive association rules. In <i>Proceedings of the 2003 SIAM International Conference on Data Mining</i>: SIAM.
[250]
Zhang, Y. &#38; Chen, X. (2018). Explainable recommendation: A survey and new perspectives. <i>arXiv preprint arXiv:1804.11192</i>.
[251]
Zhao, X., Wu, Y., Lee, D. L., &#38; Cui, W. (2019). iforest: Interpreting random forests via visual analytics. <i>IEEE transactions on visualization and computer graphics</i>.
[252]
Zhou, Z. H., Chen, S. F., &#38; Chen, Z. Q. (2000). A statistics based approach for extracting priority rules from trained neural networks. In <i>ijcnn</i>: IEEE.
[253]
Zhou, Z. H., Jiang, Y., &#38; Chen, S. F. (2003). Extracting symbolic rules from trained neural network ensembles. <i>Ai Communications</i>.
[254]
Zilke, E., Menc&#237;a, L., &#38; Janssen, F. (2016). Deepred-rule extraction from deep neural networks. In <i>International Conference on Discovery Science</i>: Springer.

Cited By

View all
  • (2025)Automatic algorithm selection for Pseudo-Boolean optimization with given computational time limitsComputers and Operations Research10.1016/j.cor.2024.106836173:COnline publication date: 1-Jan-2025
  • (2024)Assessing the brittleness of safety alignment via pruning and low-rank modificationsProceedings of the 41st International Conference on Machine Learning10.5555/3692070.3694226(52588-52610)Online publication date: 21-Jul-2024
  • (2024)Position: TRUSTLLMProceedings of the 41st International Conference on Machine Learning10.5555/3692070.3692883(20166-20270)Online publication date: 21-Jul-2024
  • Show More Cited By

Index Terms

  1. A Survey on the Explainability of Supervised Machine Learning
        Index terms have been assigned to the content through auto-classification.

        Recommendations

        Comments

        Information & Contributors

        Information

        Published In

        cover image Journal of Artificial Intelligence Research
        Journal of Artificial Intelligence Research  Volume 70, Issue
        May 2021
        1613 pages

        Publisher

        AI Access Foundation

        El Segundo, CA, United States

        Publication History

        Published: 01 May 2021
        Published in JAIR Volume 70

        Qualifiers

        • Article

        Contributors

        Other Metrics

        Bibliometrics & Citations

        Bibliometrics

        Article Metrics

        • Downloads (Last 12 months)772
        • Downloads (Last 6 weeks)75
        Reflects downloads up to 31 Jan 2025

        Other Metrics

        Citations

        Cited By

        View all
        • (2025)Automatic algorithm selection for Pseudo-Boolean optimization with given computational time limitsComputers and Operations Research10.1016/j.cor.2024.106836173:COnline publication date: 1-Jan-2025
        • (2024)Assessing the brittleness of safety alignment via pruning and low-rank modificationsProceedings of the 41st International Conference on Machine Learning10.5555/3692070.3694226(52588-52610)Online publication date: 21-Jul-2024
        • (2024)Position: TRUSTLLMProceedings of the 41st International Conference on Machine Learning10.5555/3692070.3692883(20166-20270)Online publication date: 21-Jul-2024
        • (2024)Causal Explanations for Sequential Decision-Making in Multi-Agent SystemsProceedings of the 23rd International Conference on Autonomous Agents and Multiagent Systems10.5555/3635637.3662930(771-779)Online publication date: 6-May-2024
        • (2024)The fairness fairProceedings of the Thirty-Eighth AAAI Conference on Artificial Intelligence and Thirty-Sixth Conference on Innovative Applications of Artificial Intelligence and Fourteenth Symposium on Educational Advances in Artificial Intelligence10.1609/aaai.v38i20.30272(22624-22631)Online publication date: 20-Feb-2024
        • (2024)CrystalBoxProceedings of the Thirty-Eighth AAAI Conference on Artificial Intelligence and Thirty-Sixth Conference on Innovative Applications of Artificial Intelligence and Fourteenth Symposium on Educational Advances in Artificial Intelligence10.1609/aaai.v38i13.29372(14563-14571)Online publication date: 20-Feb-2024
        • (2024)Frontiers in OperationsManufacturing & Service Operations Management10.1287/msom.2022.064126:4(1286-1305)Online publication date: 1-Jul-2024
        • (2024)Knowledge-graph-based explainable AIJournal of Information Science10.1177/0165551522111284450:4(1019-1029)Online publication date: 1-Aug-2024
        • (2024)A Survey of Machine Learning for Urban Decision Making: Applications in Planning, Transportation, and HealthcareACM Computing Surveys10.1145/369598657:4(1-41)Online publication date: 22-Nov-2024
        • (2024)From Transparency to Accountability and Back: A Discussion of Access and Evidence in AI AuditingProceedings of the 4th ACM Conference on Equity and Access in Algorithms, Mechanisms, and Optimization10.1145/3689904.3694711(1-14)Online publication date: 29-Oct-2024
        • Show More Cited By

        View Options

        View options

        PDF

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader

        Login options

        Full Access

        Figures

        Tables

        Media

        Share

        Share

        Share this Publication link

        Share on social media