Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Skip to main content

Improving Understandability of Explanations with a Usage of Expert Knowledge

  • Conference paper
  • First Online:
Artificial Intelligence. ECAI 2023 International Workshops (ECAI 2023)

Abstract

Data analysis is one of the most important parts of data mining and machine learning tasks. In recent years, explainable artificial intelligence methods have been used very often to support this phase. However, the explanations themselves are very often difficult to understand by domain experts, who play one of the most important roles in the phase of data analysis. In this work, we proposed a procedure to combine domain knowledge with ML and XAI methods to improve the understandability of explanations. We demonstrated the feasibility of our approach on a publicly available medical dataset. We describe a procedure for obtaining intuitively interpretable information about distinguishable groups of patients and defining differences between them with the usage of clustering, rule–based encoded domain knowledge, and SHAP values.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 89.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 119.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Body mass index : considerations for practitioners, August 2. https://stacks.cdc.gov/view/cdc/25368, pamphlet (or booklet)

  2. Akindele, M., Phillips, J., Igumbor, E.: The relationship between body fat percentage and body mass index in overweight and obese individuals in an urban African setting. J. Public Health Africa 7, 15–19 (2016)

    Google Scholar 

  3. Arya, V., et al.: One explanation does not fit all: a toolkit and taxonomy of AI explainability techniques. CoRR abs/1909.03012 (2019). https://arxiv.org/abs/1909.03012

  4. Breiman, L.: Statistical modeling: the two cultures (with comments and a rejoinder by the author). Stat. Sci. 16(3), 199–231 (2001)

    Article  Google Scholar 

  5. Collaris, D., van Wijk, J.J.: Explainexplore: visual exploration of machine learning explanations. In: 2020 IEEE Pacific Visualization Symposium (PacificVis), pp. 26–35 (2020)

    Google Scholar 

  6. Ćwiek-Kupczyńska, H., et al.: Semantic concept schema of the linear mixed model of experimental observations. Sci. Data 7(1), 70 (2020). https://doi.org/10.1038/s41597-020-0409-7

  7. Hastie, T., Tibshirani, R., Friedman, J., Franklin, J.: The elements of statistical learning: data mining, inference, and prediction. Math. Intell. 27, 83–85 (2004)

    Google Scholar 

  8. of Health, U.S.D., for Disease Control, H.S.C., for Health Statistics, P.N.C.: National health and nutrition examination survey (nhanes), 1999–2000 (2012). https://doi.org/10.3886/ICPSR25501.v4

  9. Holzinger, A.: Explainable AI and multi-modal causability in medicine. i-com 19(3), 171–179 (2020). https://doi.org/10.1515/icom-2020-0024

  10. Jin, W., Li, X., Hamarneh, G.: Evaluating explainable AI on a multi-modal medical imaging task: Can existing algorithms fulfill clinical requirements? In: AAAI Conference on Artificial Intelligence, March 2022

    Google Scholar 

  11. Klein, L., El-Assady, M., Jäger, P.F.: From correlation to causation: formalizing interpretable machine learning as a statistical process (2022). https://arxiv.org/abs/2207.04969

  12. Kleiser, C., Schaffrath Rosario, A., Mensink, G.B., Prinz-Langenohl, R., Kurth, B.M.: Potential determinants of obesity among children and adolescents in germany: results from the cross-sectional kiggs study. BMC Public Health 9(1), 46 (2009). https://doi.org/10.1186/1471-2458-9-46

  13. Lawrynowicz, A.: Semantic data mining: an ontology-based approach, April 2017

    Google Scholar 

  14. Li, Z., Zhang, C., Zhang, Y., Zhang, J.: Semanticaxis: exploring multi-attribute data by semantic construction and ranking analysis. J. Vis. 24(5), 1065–1081 (2021). https://doi.org/10.1007/s12650-020-00733-z

  15. Lundberg, S.M., et al.: Explainable AI for trees: From local explanations to global understanding. CoRR abs/1905.04610 (2019). https://arxiv.org/abs/1905.04610

  16. Mothilal, R.K., Sharma, A., Tan, C.: Explaining machine learning classifiers through diverse counterfactual explanations. In: Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency, ACM, January 2020. https://doi.org/10.1145/3351095.3372850

  17. Nalepa, G.J., Bobek, S., Kutt, K., Atzmueller, M.: Semantic data mining in ubiquitous sensing: a survey. Sensors 21(13) (2021). https://www.mdpi.com/1424-8220/21/13/4322

  18. Novosad, S., Khan, S., Wolfe, B.N., Khan, A.: Role of obesity in asthma control, the obesity-asthma phenotype. J. Allergy 2013, 538642 (2013)

    Article  Google Scholar 

  19. Ribeiro, M.T., Singh, S., Guestrin, C.: Anchors: high-precision model-agnostic explanations. In: AAAI Conference on Artificial Intelligence (2018)

    Google Scholar 

  20. Sirichanya, C., Kraisak, K.: Semantic data mining in the information age: a systematic review. Int. J. Intell. Syst. 36(8), 3880–3916 (2021). https://onlinelibrary.wiley.com/doi/abs/10.1002/int.22443

  21. Smith, J.W., Everhart, J.E., Dickson, W., Knowler, W.C., Johannes, R.S.: Using the adap learning algorithm to forecast the onset of diabetes mellitus. In: Proceedings of Annual Symposium on Computer Application in Medical Care, p. 261. American Medical Informatics Ass. (1988)

    Google Scholar 

  22. Suchanek, F., Weikum, G.: Knowledge bases in the age of big data analytics. Proc. VLDB Endowment 7, 1713–1714 (2014)

    Google Scholar 

  23. Visani, G., Bagli, E., Chesani, F.: Optilime: optimized LIME explanations for diagnostic computer algorithms. CoRR abs/2006.05714 (2020). https://arxiv.org/abs/2006.05714

Download references

Acknowledgements

This paper is funded from the XPM (Explainable Predictive Maintenance) project funded by the National Science Center, Poland under CHIST-ERA programme Grant Agreement No. 857925 (NCN UMO-2020/02/Y/ST6/00070).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Maciej Szelążek .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Szelążek, M., Bobek, S., Nalepa, G.J. (2024). Improving Understandability of Explanations with a Usage of Expert Knowledge. In: Nowaczyk, S., et al. Artificial Intelligence. ECAI 2023 International Workshops. ECAI 2023. Communications in Computer and Information Science, vol 1948. Springer, Cham. https://doi.org/10.1007/978-3-031-50485-3_3

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-50485-3_3

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-50484-6

  • Online ISBN: 978-3-031-50485-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics