Please refer to the course Webpage for updated information:
https://human-centered.ai/lv-185-a83-machine-learning-for-health-informatics-class-of-2021/
Medicine is evolving into a data-driven science. Health AI is working to effectively and efficiently use machine learning methods to solve problems in the comprehensive field of health and life sciences. This master's course takes a research-centered teaching approach. Topics covered include methods for combining human intelligence and machine intelligence to support medical decision making. Since 2018, the European General Data Protection Regulation explicitly provides for a legal "right to explanation", and the EU Parliament recently adopted a resolution on "explainable AI" as part of the European Digitization Initiative. This calls for solutions that must enable medical experts to understand, replicate and comprehend machine results. The central focus of Class 2021 is even more on making machine decisions transparent, comprehensible and interpretable for medical experts. A critical requirement for successful AI applications in the future will be that human experts must be able to at least understand the context and be able to explore the underlying explanatory factors, with the goal of answering the question WHY a particular machine decision was made. This is desirable in many domains, but mandatory in the medical domain. In addition, explainable AI should enable a healthcare professional to ask counterfactual questions, such as "what if?" questions, to also gain new insights. Ultimately, such approaches foster confidence for future solutions from artificial intelligence - which will inevitably enter everyday medical practice.
For further questions please ask directly the course director: Andreas Holzinger
ECTS-Breakdown (sum=75h, corresponds with 3 ECTS, where 1 ECTS = 25 h students workload):
a) Presence during the lecture 8 x 3 h = 24 h
b) Preparation before and after the lecture 8 x 1 h = 8 h
c) Preparation of assignments and presentation 28 h + 2 h = 30 h
d) Written exam including exam preparation 1 h + 12 h = 13 h
Sum TOTAL student's workload = 75 h
Holzinger, A. 2014. Biomedical Informatics: Discovering Knowledge in Big Data, New York, Springer, doi:10.1007/978-3-319-04528-3.
Holzinger, A. (ed.) 2016. Machine Learning for Health Informatics: State-of-the-Art and Future Challenges, Lecture Notes in Artificial Intelligence LNAI 9605, Cham: Springer International, doi:10.1007/978-3-319-50478-0.
Holzinger, A. 2016. Interactive Machine Learning for Health Informatics: When do we need the human-in-the-loop? Brain Informatics, 3, (2), 119-131, doi:10.1007/s40708-016-0042-6.
Interest in machine learning with application to health informatics with a special focus on privacy, security, data protection safety, ethical and social issues and in explainable AI [4] and the doctor-in-the-loop [5], which led to the developement of the concept of Cau-sa-bility (in accordance to Usa-bi-lity) to evaluate the quality of explanations [6, 7, 8]
[4] Andreas Holzinger (2018): From Machine Learning to Explainable AI. 2018 World Symposium on Digital Intelligence for Systems and Machines (DISA), 23-25 Aug. 2018 2018. 55-66, doi:10.1109/DISA.2018.8490530.
[5] Andreas Holzinger (2016): Interactive Machine Learning for Health Informatics: When do we need the human-in-the-loop? Brain Informatics, 3, (2), 119-131, doi:10.1007/s40708-016-0042-6
[6] Andreas Holzinger, Georg Langs, Helmut Denk, Kurt Zatloukal & Heimo Müller (2019). Causability and Explainability of Artificial Intelligence in Medicine. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery, 9, (4), 1-13, doi:10.1002/widm.1312
[7] Andreas Holzinger, Andre Carrington & Heimo Müller (2020). Measuring the Quality of Explanations: The System Causability Scale (SCS). Comparing Human and Machine Explanations. KI - Künstliche Intelligenz (German Journal of Artificial intelligence), Special Issue on Interactive Machine Learning, Edited by Kristian Kersting, TU Darmstadt, 34, (2), 193-198, doi:10.1007/s13218-020-00636-z
[8] Andreas Holzinger, Bernd Malle, Anna Saranti & Bastian Pfeifer (2021). Towards Multi-Modal Causability with Graph Neural Networks enabling Information Fusion for explainable AI. Information Fusion, 71, (7), 28-37, doi:10.1016/j.inffus.2021.01.008
https://www.sciencedirect.com/science/article/pii/S1566253521000142?dgcid=rss_sd_all
More information on the Course Homepage!