Show simple item record

Files in this item

Thumbnail

Item metadata

dc.contributor.authorInuwa-Dutse, Isa
dc.contributor.authorToniolo, Alice
dc.contributor.authorWeller, Adrian
dc.contributor.authorBhatt, Umang
dc.date.accessioned2023-11-28T09:30:06Z
dc.date.available2023-11-28T09:30:06Z
dc.date.issued2023-12
dc.identifier296886418
dc.identifierd24bca36-ea02-4f4a-9b0f-96462c771d36
dc.identifier.citationInuwa-Dutse , I , Toniolo , A , Weller , A & Bhatt , U 2023 , ' Algorithmic loafing and mitigation strategies in Human-AI teams ' , Computers in Human Behavior: Artificial Humans , vol. 1 , no. 2 , 100024 . https://doi.org/10.1016/j.chbah.2023.100024en
dc.identifier.issn2949-8821
dc.identifier.otherJisc: 1486182
dc.identifier.otherORCID: /0000-0002-6816-6360/work/147967068
dc.identifier.urihttps://hdl.handle.net/10023/28772
dc.descriptionThis research work was initiated under the Scottish Informatics & Computer Alliance (SICSA) Remote Collaboration Activities when the first author was working at the University of St Andrews, UK. We would like to thank the SICSA for the partial funding of the research work.en
dc.description.abstractExercising social loafing – exerting minimal effort by an individual in a group setting – in human-machine teams could critically degrade performance, especially in high-stakes domains where human judgement is essential. Akin to social loafing in human interaction, algorithmic loafing may occur when humans mindlessly adhere to machine recommendations due to reluctance to engage analytically with AI recommendations and explanations. We consider how algorithmic loafing could emerge and how to mitigate it. Specifically, we posit that algorithmic loafing can be induced through repeated encounters with correct decisions from the AI and transparency may combat it. As a form of transparency, explanation is offered for reasons that include justification, control, and discovery. However, algorithmic loafing is further reinforced by the perceived competence that an explanation provides. In this work, we explored these ideas via human subject experiments (n = 239). We also study how improving decision transparency through validation by an external human approver affects performance. Using eight experimental conditions in a high-stakes criminal justice context, we find that decision accuracy is typically unaffected by multiple forms of transparency but there is a significant difference in performance when the machine errs. Participants who saw explanations alone are better at overriding incorrect decisions; however, those under induced algorithmic loafing exhibit poor performance with variation in decision time. We conclude with recommendations on curtailing algorithmic loafing and achieving social facilitation, where task visibility motivates individuals to perform better.
dc.format.extent14
dc.format.extent1557300
dc.language.isoeng
dc.relation.ispartofComputers in Human Behavior: Artificial Humansen
dc.subjectExplainable AIen
dc.subjectSocial loafingen
dc.subjectTransparent AIen
dc.subjectAlgorithmic appreciationen
dc.subjectAlgorithmic loafingen
dc.subjectQA75 Electronic computers. Computer scienceen
dc.subjectNDASen
dc.subjectSDG 16 - Peace, Justice and Strong Institutionsen
dc.subject.lccQA75en
dc.titleAlgorithmic loafing and mitigation strategies in Human-AI teamsen
dc.typeJournal articleen
dc.contributor.institutionUniversity of St Andrews. School of Computer Scienceen
dc.identifier.doi10.1016/j.chbah.2023.100024
dc.description.statusPeer revieweden


This item appears in the following Collection(s)

Show simple item record