TY - GEN
T1 - Nudging through Friction
T2 - 8th IEEE International Conference on Behavioural and Social Computing, BESC 2021
AU - Naiseh, Mohammad
AU - Al-Mansoori, Reem S.
AU - Al-Thani, Dena
AU - Jiang, Nan
AU - Ali, Raian
N1 - Publisher Copyright:
© 2021 IEEE
PY - 2021
Y1 - 2021
N2 - Explainability has become an essential requirement for safe and effective collaborative Human-AI environments, especially when generating recommendations through black-box modality. One goal of eXplainable AI (XAI) is to help humans calibrate their trust while working with intelligent systems, i.e., avoid situations where human decision-makers over-trust the AI when it is incorrect, or under-trust the AI when it is correct. XAI, in this context, aims to help humans understand AI reasoning and decide whether to follow or reject its recommendations. However, recent studies showed that users, on average, continue to overtrust (or under-trust) AI recommendations which is an indication of XAI's failure to support trust calibration. Such a failure to aid trust calibration was due to the assumption that XAI users would cognitively engage with explanations and interpret them without bias. In this work, we hypothesize that XAI interaction design can play a role in helping users' cognitive engagement with XAI and consequently enhance trust calibration. To this end, we propose friction as a Nudge-based approach to help XAI users to calibrate their trust in AI and present the results of a preliminary study of its potential in fulfilling that role.
AB - Explainability has become an essential requirement for safe and effective collaborative Human-AI environments, especially when generating recommendations through black-box modality. One goal of eXplainable AI (XAI) is to help humans calibrate their trust while working with intelligent systems, i.e., avoid situations where human decision-makers over-trust the AI when it is incorrect, or under-trust the AI when it is correct. XAI, in this context, aims to help humans understand AI reasoning and decide whether to follow or reject its recommendations. However, recent studies showed that users, on average, continue to overtrust (or under-trust) AI recommendations which is an indication of XAI's failure to support trust calibration. Such a failure to aid trust calibration was due to the assumption that XAI users would cognitively engage with explanations and interpret them without bias. In this work, we hypothesize that XAI interaction design can play a role in helping users' cognitive engagement with XAI and consequently enhance trust calibration. To this end, we propose friction as a Nudge-based approach to help XAI users to calibrate their trust in AI and present the results of a preliminary study of its potential in fulfilling that role.
KW - Calibrated trust
KW - Digital nudging
KW - Explainable AI
KW - Friction
KW - Human-AI interaction
UR - http://www.scopus.com/inward/record.url?scp=85124012774&partnerID=8YFLogxK
U2 - 10.1109/BESC53957.2021.9635271
DO - 10.1109/BESC53957.2021.9635271
M3 - Conference contribution
AN - SCOPUS:85124012774
T3 - Proceedings of 2021 8th IEEE International Conference on Behavioural and Social Computing, BESC 2021
BT - Proceedings of 2021 8th IEEE International Conference on Behavioural and Social Computing, BESC 2021
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 29 October 2021 through 31 October 2021
ER -