AI trust: Can explainable AI enhance warranted trust?

Detalhes bibliográficos
Autor(a) principal: Duarte, R.
Data de Publicação: 2023
Outros Autores: Correia, F., Arriaga, P., Paiva, A.
Tipo de documento: Artigo
Idioma: eng
Título da fonte: Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos)
Texto Completo: http://hdl.handle.net/10071/29696
Resumo: Explainable artificial intelligence (XAI), known to produce explanations so that predictions from AI models can be understood, is commonly used to mitigate possible AI mistrust. The underlying premise is that the explanations of the XAI models enhance AI trust. However, such an increase may depend on many factors. This article examined how trust in an AI recommendation system is affected by the presence of explanations, the performance of the system, and the level of risk. Our experimental study, conducted with 215 participants, has shown that the presence of explanations increases AI trust, but only in certain conditions. AI trust was higher when explanations with feature importance were provided than with counterfactual explanations. Moreover, when the system performance is not guaranteed, the use of explanations seems to lead to an overreliance on the system. Lastly, system performance had a stronger impact on trust, compared to the effects of other factors (explanation and risk).
id RCAP_0bbb32b43c15fb4045ee731322c56167
oai_identifier_str oai:repositorio.iscte-iul.pt:10071/29696
network_acronym_str RCAP
network_name_str Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos)
repository_id_str 7160
spelling AI trust: Can explainable AI enhance warranted trust?Artificial intelligence (AI)TrustExplainable AIRiskExplainable artificial intelligence (XAI), known to produce explanations so that predictions from AI models can be understood, is commonly used to mitigate possible AI mistrust. The underlying premise is that the explanations of the XAI models enhance AI trust. However, such an increase may depend on many factors. This article examined how trust in an AI recommendation system is affected by the presence of explanations, the performance of the system, and the level of risk. Our experimental study, conducted with 215 participants, has shown that the presence of explanations increases AI trust, but only in certain conditions. AI trust was higher when explanations with feature importance were provided than with counterfactual explanations. Moreover, when the system performance is not guaranteed, the use of explanations seems to lead to an overreliance on the system. Lastly, system performance had a stronger impact on trust, compared to the effects of other factors (explanation and risk).Wiley2023-11-21T15:16:16Z2023-01-01T00:00:00Z20232023-11-21T15:15:24Zinfo:eu-repo/semantics/publishedVersioninfo:eu-repo/semantics/articleapplication/pdfhttp://hdl.handle.net/10071/29696eng2578-186310.1155/2023/4637678Duarte, R.Correia, F.Arriaga, P.Paiva, A.info:eu-repo/semantics/openAccessreponame:Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos)instname:Agência para a Sociedade do Conhecimento (UMIC) - FCT - Sociedade da Informaçãoinstacron:RCAAP2023-11-26T01:16:56Zoai:repositorio.iscte-iul.pt:10071/29696Portal AgregadorONGhttps://www.rcaap.pt/oai/openaireopendoar:71602024-03-19T23:19:43.550265Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos) - Agência para a Sociedade do Conhecimento (UMIC) - FCT - Sociedade da Informaçãofalse
dc.title.none.fl_str_mv AI trust: Can explainable AI enhance warranted trust?
title AI trust: Can explainable AI enhance warranted trust?
spellingShingle AI trust: Can explainable AI enhance warranted trust?
Duarte, R.
Artificial intelligence (AI)
Trust
Explainable AI
Risk
title_short AI trust: Can explainable AI enhance warranted trust?
title_full AI trust: Can explainable AI enhance warranted trust?
title_fullStr AI trust: Can explainable AI enhance warranted trust?
title_full_unstemmed AI trust: Can explainable AI enhance warranted trust?
title_sort AI trust: Can explainable AI enhance warranted trust?
author Duarte, R.
author_facet Duarte, R.
Correia, F.
Arriaga, P.
Paiva, A.
author_role author
author2 Correia, F.
Arriaga, P.
Paiva, A.
author2_role author
author
author
dc.contributor.author.fl_str_mv Duarte, R.
Correia, F.
Arriaga, P.
Paiva, A.
dc.subject.por.fl_str_mv Artificial intelligence (AI)
Trust
Explainable AI
Risk
topic Artificial intelligence (AI)
Trust
Explainable AI
Risk
description Explainable artificial intelligence (XAI), known to produce explanations so that predictions from AI models can be understood, is commonly used to mitigate possible AI mistrust. The underlying premise is that the explanations of the XAI models enhance AI trust. However, such an increase may depend on many factors. This article examined how trust in an AI recommendation system is affected by the presence of explanations, the performance of the system, and the level of risk. Our experimental study, conducted with 215 participants, has shown that the presence of explanations increases AI trust, but only in certain conditions. AI trust was higher when explanations with feature importance were provided than with counterfactual explanations. Moreover, when the system performance is not guaranteed, the use of explanations seems to lead to an overreliance on the system. Lastly, system performance had a stronger impact on trust, compared to the effects of other factors (explanation and risk).
publishDate 2023
dc.date.none.fl_str_mv 2023-11-21T15:16:16Z
2023-01-01T00:00:00Z
2023
2023-11-21T15:15:24Z
dc.type.status.fl_str_mv info:eu-repo/semantics/publishedVersion
dc.type.driver.fl_str_mv info:eu-repo/semantics/article
format article
status_str publishedVersion
dc.identifier.uri.fl_str_mv http://hdl.handle.net/10071/29696
url http://hdl.handle.net/10071/29696
dc.language.iso.fl_str_mv eng
language eng
dc.relation.none.fl_str_mv 2578-1863
10.1155/2023/4637678
dc.rights.driver.fl_str_mv info:eu-repo/semantics/openAccess
eu_rights_str_mv openAccess
dc.format.none.fl_str_mv application/pdf
dc.publisher.none.fl_str_mv Wiley
publisher.none.fl_str_mv Wiley
dc.source.none.fl_str_mv reponame:Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos)
instname:Agência para a Sociedade do Conhecimento (UMIC) - FCT - Sociedade da Informação
instacron:RCAAP
instname_str Agência para a Sociedade do Conhecimento (UMIC) - FCT - Sociedade da Informação
instacron_str RCAAP
institution RCAAP
reponame_str Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos)
collection Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos)
repository.name.fl_str_mv Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos) - Agência para a Sociedade do Conhecimento (UMIC) - FCT - Sociedade da Informação
repository.mail.fl_str_mv
_version_ 1799135496250589184