Explainable Recommendations in Intelligent Systems: Delivery Methods, Modalities and Risks
Authors: Naiseh, M., Jiang, N., Ma, J. and Ali, R.
Journal: Lecture Notes in Business Information Processing
Volume: 385 LNBIP
Pages: 212-228
eISSN: 1865-1356
ISSN: 1865-1348
DOI: 10.1007/978-3-030-50316-1_13
Abstract:With the increase in data volume, velocity and types, intelligent human-agent systems have become popular and adopted in different application domains, including critical and sensitive areas such as health and security. Humans’ trust, their consent and receptiveness to recommendations are the main requirement for the success of such services. Recently, the demand on explaining the recommendations to humans has increased both from humans interacting with these systems so that they make an informed decision and, also, owners and systems managers to increase transparency and consequently trust and users’ retention. Existing systematic reviews in the area of explainable recommendations focused on the goal of providing explanations, their presentation and informational content. In this paper, we review the literature with a focus on two user experience facets of explanations; delivery methods and modalities. We then focus on the risks of explanation both on user experience and their decision making. Our review revealed that explanations delivery to end-users is mostly designed to be along with the recommendation in a push and pull styles while archiving explanations for later accountability and traceability is still limited. We also found that the emphasis was mainly on the benefits of recommendations while risks and potential concerns, such as over-reliance on machines, is still a new area to explore.
https://eprints.bournemouth.ac.uk/34312/
Source: Scopus
Explainable Recommendations in Intelligent Systems: Delivery Methods, Modalities and Risks
Authors: Naiseh, M., Jiang, N., Ma, J. and Ali, R.
Journal: RESEARCH CHALLENGES IN INFORMATION SCIENCE (RCIS 2020)
Volume: 385
Pages: 212-228
eISSN: 1865-1356
ISBN: 978-3-030-50315-4
ISSN: 1865-1348
DOI: 10.1007/978-3-030-50316-1_13
https://eprints.bournemouth.ac.uk/34312/
Source: Web of Science (Lite)
Explainable Recommendations in Intelligent Systems: Delivery Methods, Modalities and Risks
Authors: Naiseh, M., Jiang, N., Ma, J. and Ali, R.
Conference: RCIS 2020: Research Challenges in Information Science
Pages: 212-228
ISBN: 9783030503154
ISSN: 1865-1348
Abstract:With the increase in data volume, velocity and types, intelligent human-agent systems have become popular and adopted in different application domains, including critical and sensitive areas such as health and security. Humans’ trust, their consent and receptiveness to recommendations are the main requirement for the success of such services. Recently, the demand on explaining the recommendations to humans has increased both from humans interacting with these systems so that they make an informed decision and, also, owners and systems managers to increase transparency and consequently trust and users’ retention. Existing systematic reviews in the area of explainable recommendations focused on the goal of providing explanations, their presentation and informational content. In this paper, we review the literature with a focus on two user experience facets of explanations; delivery methods and modalities. We then focus on the risks of explanation both on user experience and their decision making. Our review revealed that explanations delivery to end-users is mostly designed to be along with the recommendation in a push and pull styles while archiving explanations for later accountability and traceability is still limited. We also found that the emphasis was mainly on the benefits of recommendations while risks and potential concerns, such as over-reliance on machines, is still a new area to explore.
https://eprints.bournemouth.ac.uk/34312/
Source: BURO EPrints