More Is Not Always Better : Impacts of AI-Generated Confidence and Explanations in Human-Automation Interaction
OBJECTIVE: The study aimed to enhance transparency in autonomous systems by automatically generating and visualizing confidence and explanations and assessing their impacts on performance, trust, preference, and eye-tracking behaviors in human-automation interaction.
BACKGROUND: System transparency is vital to maintaining appropriate levels of trust and mission success. Previous studies presented mixed results regarding the impact of displaying likelihood information and explanations, and often relied on hand-created information, limiting scalability and failing to address real-world dynamics.
METHOD: We conducted a dual-task experiment involving 42 university students who operated a simulated surveillance testbed with assistance from intelligent detectors. The study used a 2 (confidence visualization: yes vs. no) × 3 (visual explanations: none, bounding boxes, bounding boxes and keypoints) mixed design. Task performance, human trust, preference for intelligent detectors, and eye-tracking behaviors were evaluated.
RESULTS: Visual explanations using bounding boxes and keypoints improved detection task performance when confidence was not displayed. Meanwhile, visual explanations enhanced trust and preference for the intelligent detector, regardless of the explanation type. Confidence visualization did not influence human trust in and preference for the intelligent detector. Moreover, both visual information slowed saccade velocities.
CONCLUSION: The study demonstrated that visual explanations could improve performance, trust, and preference in human-automation interaction without confidence visualization partially by changing the search strategies. However, excessive information might cause adverse effects.
APPLICATION: These findings provide guidance for the design of transparent automation, emphasizing the importance of context-appropriate and user-centered explanations to foster effective human-machine collaboration.
Medienart: |
E-Artikel |
---|
Erscheinungsjahr: |
2024 |
---|---|
Erschienen: |
2024 |
Enthalten in: |
Zur Gesamtaufnahme - year:2024 |
---|---|
Enthalten in: |
Human factors - (2024) vom: 04. März, Seite 187208241234810 |
Sprache: |
Englisch |
---|
Beteiligte Personen: |
Ling, Shihong [VerfasserIn] |
---|
Links: |
---|
Themen: |
Explainable artificial intelligence |
---|
Anmerkungen: |
Date Revised 04.03.2024 published: Print-Electronic Citation Status Publisher |
---|
doi: |
10.1177/00187208241234810 |
---|
funding: |
|
---|---|
Förderinstitution / Projekttitel: |
|
PPN (Katalog-ID): |
NLM36927461X |
---|
LEADER | 01000naa a22002652 4500 | ||
---|---|---|---|
001 | NLM36927461X | ||
003 | DE-627 | ||
005 | 20240305232906.0 | ||
007 | cr uuu---uuuuu | ||
008 | 240305s2024 xx |||||o 00| ||eng c | ||
024 | 7 | |a 10.1177/00187208241234810 |2 doi | |
028 | 5 | 2 | |a pubmed24n1317.xml |
035 | |a (DE-627)NLM36927461X | ||
035 | |a (NLM)38437598 | ||
040 | |a DE-627 |b ger |c DE-627 |e rakwb | ||
041 | |a eng | ||
100 | 1 | |a Ling, Shihong |e verfasserin |4 aut | |
245 | 1 | 0 | |a More Is Not Always Better |b Impacts of AI-Generated Confidence and Explanations in Human-Automation Interaction |
264 | 1 | |c 2024 | |
336 | |a Text |b txt |2 rdacontent | ||
337 | |a ƒaComputermedien |b c |2 rdamedia | ||
338 | |a ƒa Online-Ressource |b cr |2 rdacarrier | ||
500 | |a Date Revised 04.03.2024 | ||
500 | |a published: Print-Electronic | ||
500 | |a Citation Status Publisher | ||
520 | |a OBJECTIVE: The study aimed to enhance transparency in autonomous systems by automatically generating and visualizing confidence and explanations and assessing their impacts on performance, trust, preference, and eye-tracking behaviors in human-automation interaction | ||
520 | |a BACKGROUND: System transparency is vital to maintaining appropriate levels of trust and mission success. Previous studies presented mixed results regarding the impact of displaying likelihood information and explanations, and often relied on hand-created information, limiting scalability and failing to address real-world dynamics | ||
520 | |a METHOD: We conducted a dual-task experiment involving 42 university students who operated a simulated surveillance testbed with assistance from intelligent detectors. The study used a 2 (confidence visualization: yes vs. no) × 3 (visual explanations: none, bounding boxes, bounding boxes and keypoints) mixed design. Task performance, human trust, preference for intelligent detectors, and eye-tracking behaviors were evaluated | ||
520 | |a RESULTS: Visual explanations using bounding boxes and keypoints improved detection task performance when confidence was not displayed. Meanwhile, visual explanations enhanced trust and preference for the intelligent detector, regardless of the explanation type. Confidence visualization did not influence human trust in and preference for the intelligent detector. Moreover, both visual information slowed saccade velocities | ||
520 | |a CONCLUSION: The study demonstrated that visual explanations could improve performance, trust, and preference in human-automation interaction without confidence visualization partially by changing the search strategies. However, excessive information might cause adverse effects | ||
520 | |a APPLICATION: These findings provide guidance for the design of transparent automation, emphasizing the importance of context-appropriate and user-centered explanations to foster effective human-machine collaboration | ||
650 | 4 | |a Journal Article | |
650 | 4 | |a explainable artificial intelligence | |
650 | 4 | |a eye-tracking analysis | |
650 | 4 | |a human–automation interaction | |
650 | 4 | |a task performance | |
650 | 4 | |a transparency | |
650 | 4 | |a trust | |
700 | 1 | |a Zhang, Yutong |e verfasserin |4 aut | |
700 | 1 | |a Du, Na |e verfasserin |4 aut | |
773 | 0 | 8 | |i Enthalten in |t Human factors |d 1962 |g (2024) vom: 04. März, Seite 187208241234810 |w (DE-627)NLM000220442 |x 1547-8181 |7 nnns |
773 | 1 | 8 | |g year:2024 |g day:04 |g month:03 |g pages:187208241234810 |
856 | 4 | 0 | |u http://dx.doi.org/10.1177/00187208241234810 |3 Volltext |
912 | |a GBV_USEFLAG_A | ||
912 | |a GBV_NLM | ||
951 | |a AR | ||
952 | |j 2024 |b 04 |c 03 |h 187208241234810 |