Abstract
Explainable AI (XAI) is increasingly being used in the healthcare domain. In health management, clinicians and patients are critical stakeholders, requiring tailored XAI explanations based on their unique needs. Our study investigates the differences in explanation needs between clinicians and patients and designs corresponding explanation interfaces for each group. Using a scenario-based approach, we assessed stakeholder-tailored needs, analyzed differences, and designed interfaces using theoretical frameworks. The results demonstrate diverse stakeholder motivations for seeking explanations, leading to varied requirements. The designed interfaces effectively address these requirements, as validated by the preference selection and qualitative feedback from clinicians and patients. Their suggestions provide design insights and highlight the divergent needs of these stakeholder groups. This study contributes practical and theoretical implications to XAI research, emphasizing the importance of understanding diverse stakeholder needs and incorporating relevant theoretical concepts into user-centered interface design.
| Original language | English |
|---|---|
| Article number | 103160 |
| Journal | International Journal of Human Computer Studies |
| Volume | 181 |
| DOIs | |
| State | Published - Jan 2024 |
Bibliographical note
Publisher Copyright:© 2023
UN SDGs
This output contributes to the following UN Sustainable Development Goals (SDGs)
-
SDG 3 Good Health and Well-being
Keywords
- Digital health
- Explanation interfaces
- Explanation needs
- Health management
- Human-centered XAI
- Medical XAI
Fingerprint
Dive into the research topics of 'Do stakeholder needs differ? - Designing stakeholder-tailored Explainable Artificial Intelligence (XAI) interfaces'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver