Files in this item
Algorithmic loafing and mitigation strategies in Human-AI teams
Item metadata
dc.contributor.author | Inuwa-Dutse, Isa | |
dc.contributor.author | Toniolo, Alice | |
dc.contributor.author | Weller, Adrian | |
dc.contributor.author | Bhatt, Umang | |
dc.date.accessioned | 2023-11-28T09:30:06Z | |
dc.date.available | 2023-11-28T09:30:06Z | |
dc.date.issued | 2023-12 | |
dc.identifier | 296886418 | |
dc.identifier | d24bca36-ea02-4f4a-9b0f-96462c771d36 | |
dc.identifier.citation | Inuwa-Dutse , I , Toniolo , A , Weller , A & Bhatt , U 2023 , ' Algorithmic loafing and mitigation strategies in Human-AI teams ' , Computers in Human Behavior: Artificial Humans , vol. 1 , no. 2 , 100024 . https://doi.org/10.1016/j.chbah.2023.100024 | en |
dc.identifier.issn | 2949-8821 | |
dc.identifier.other | Jisc: 1486182 | |
dc.identifier.other | ORCID: /0000-0002-6816-6360/work/147967068 | |
dc.identifier.uri | https://hdl.handle.net/10023/28772 | |
dc.description | This research work was initiated under the Scottish Informatics & Computer Alliance (SICSA) Remote Collaboration Activities when the first author was working at the University of St Andrews, UK. We would like to thank the SICSA for the partial funding of the research work. | en |
dc.description.abstract | Exercising social loafing – exerting minimal effort by an individual in a group setting – in human-machine teams could critically degrade performance, especially in high-stakes domains where human judgement is essential. Akin to social loafing in human interaction, algorithmic loafing may occur when humans mindlessly adhere to machine recommendations due to reluctance to engage analytically with AI recommendations and explanations. We consider how algorithmic loafing could emerge and how to mitigate it. Specifically, we posit that algorithmic loafing can be induced through repeated encounters with correct decisions from the AI and transparency may combat it. As a form of transparency, explanation is offered for reasons that include justification, control, and discovery. However, algorithmic loafing is further reinforced by the perceived competence that an explanation provides. In this work, we explored these ideas via human subject experiments (n = 239). We also study how improving decision transparency through validation by an external human approver affects performance. Using eight experimental conditions in a high-stakes criminal justice context, we find that decision accuracy is typically unaffected by multiple forms of transparency but there is a significant difference in performance when the machine errs. Participants who saw explanations alone are better at overriding incorrect decisions; however, those under induced algorithmic loafing exhibit poor performance with variation in decision time. We conclude with recommendations on curtailing algorithmic loafing and achieving social facilitation, where task visibility motivates individuals to perform better. | |
dc.format.extent | 14 | |
dc.format.extent | 1557300 | |
dc.language.iso | eng | |
dc.relation.ispartof | Computers in Human Behavior: Artificial Humans | en |
dc.subject | Explainable AI | en |
dc.subject | Social loafing | en |
dc.subject | Transparent AI | en |
dc.subject | Algorithmic appreciation | en |
dc.subject | Algorithmic loafing | en |
dc.subject | QA75 Electronic computers. Computer science | en |
dc.subject | NDAS | en |
dc.subject | SDG 16 - Peace, Justice and Strong Institutions | en |
dc.subject.lcc | QA75 | en |
dc.title | Algorithmic loafing and mitigation strategies in Human-AI teams | en |
dc.type | Journal article | en |
dc.contributor.institution | University of St Andrews. School of Computer Science | en |
dc.identifier.doi | 10.1016/j.chbah.2023.100024 | |
dc.description.status | Peer reviewed | en |
This item appears in the following Collection(s)
Items in the St Andrews Research Repository are protected by copyright, with all rights reserved, unless otherwise indicated.