Autor: |
Federico Cabitza, Andrea Campagner, Lorenzo Famiglini, Enrico Gallazzi, Giovanni Andrea La Maida |
Přispěvatelé: |
Holzinger, A, Kieseberg, P, Min Tjoa, A, Weippl, E, Cabitza, F, Campagner, A, Famiglini, L, Gallazzi, E, La Maida, G |
Jazyk: |
angličtina |
Rok vydání: |
2022 |
Předmět: |
|
Zdroj: |
Lecture Notes in Computer Science ISBN: 9783031144622 |
Popis: |
Although deep learning-based AI systems for diagnostic imaging tasks have virtually showed superhuman accuracy, their use in medical settings has been questioned due to their “black box”, not interpretable nature. To address this shortcoming, several methods have been proposed to make AI eXplainable (XAI), including Pixel Attribution Methods; however, it is still unclear whether these methods are actually effective in “opening” the black-box and improving diagnosis, particularly in tasks where pathological conditions are difficult to detect. In this study, we focus on the detection of thoraco-lumbar fractures from X-rays with the goal of assessing the impact of PAMs on diagnostic decision making by addressing two separate research questions: first, whether activation maps (as an instance of PAM) were perceived as useful in the aforementioned task; and, second, whether maps were also capable to reduce the diagnostic error rate. We show that, even though AMs were not considered significantly useful by physicians, the image readers found high value in the maps in relation to other perceptual dimensions (i.e., pertinency, coherence) and, most importantly, their accuracy significantly improved when given XAI support in a pilot study involving 7 doctors in the interpretation of a small, but carefully chosen, set of images. |
Databáze: |
OpenAIRE |
Externí odkaz: |
|