Explainable Recommendations in Intelligent Systems: Delivery Methods, Modalities and Risks

This data was imported from Scopus:

Authors: Naiseh, M., Jiang, N., Ma, J. and Ali, R.

http://eprints.bournemouth.ac.uk/34312/

Journal: Lecture Notes in Business Information Processing

Volume: 385 LNBIP

Pages: 212-228

eISSN: 1865-1356

ISBN: 9783030503154

ISSN: 1865-1348

DOI: 10.1007/978-3-030-50316-1_13

© 2020, Springer Nature Switzerland AG. With the increase in data volume, velocity and types, intelligent human-agent systems have become popular and adopted in different application domains, including critical and sensitive areas such as health and security. Humans’ trust, their consent and receptiveness to recommendations are the main requirement for the success of such services. Recently, the demand on explaining the recommendations to humans has increased both from humans interacting with these systems so that they make an informed decision and, also, owners and systems managers to increase transparency and consequently trust and users’ retention. Existing systematic reviews in the area of explainable recommendations focused on the goal of providing explanations, their presentation and informational content. In this paper, we review the literature with a focus on two user experience facets of explanations; delivery methods and modalities. We then focus on the risks of explanation both on user experience and their decision making. Our review revealed that explanations delivery to end-users is mostly designed to be along with the recommendation in a push and pull styles while archiving explanations for later accountability and traceability is still limited. We also found that the emphasis was mainly on the benefits of recommendations while risks and potential concerns, such as over-reliance on machines, is still a new area to explore.

The data on this page was last updated at 05:24 on October 24, 2020.