Skip to main content

Explainable recommendations and calibrated trust: two systematic users’ errors.

Naiseh, M., Cemiloglu, D., Jiang, N., Althani, D. and Ali, R., 2021. Explainable recommendations and calibrated trust: two systematic users’ errors. Computer. (In Press)

Full text available as:

[img] PDF
Final version- Special Issue paper.pdf - Accepted Version
Restricted to Repository staff only
Available under License Creative Commons Attribution Non-commercial.

462kB

Abstract

The increased adoption of collaborative Human-AI decision-making tools triggered a need to explain the recommendations for safe and effective collaboration. However, evidence from the recent literature showed that current implementation of AI explanations is failing to achieve adequate trust calibration. Such failure has lead decision-makers to either end-up with over-trust, e.g., people follow incorrect recommendations or under-trust, they reject a correct recommendation. In this paper, we explore how users interact with explanations and why trust calibration errors occur. We take clinical decision-support systems as a case study. Our empirical investigation is based on think-aloud protocol and observations, supported by scenarios and decision-making exercise utilizing a set of explainable recommendations interfaces. Our study involved 16 participants from medical domain who use clinical decision support systems frequently. Our findings showed that participants had two systematic errors while interacting with the explanations either by skipping them or misapplying them in their task.

Item Type:Article
ISSN:0018-9162
Group:Faculty of Science & Technology
ID Code:35465
Deposited By: Unnamed user with email symplectic@symplectic
Deposited On:10 May 2021 10:36
Last Modified:27 May 2021 07:53

Downloads

Downloads per month over past year

More statistics for this item...
Repository Staff Only -