Journal
JOURNAL OF COGNITIVE ENGINEERING AND DECISION MAKING
Volume 16, Issue 2, Pages 101-118Publisher
SAGE PUBLICATIONS INC
DOI: 10.1177/15553434221096261
Keywords
trust; trust in automation; trust assessment; practitioner
Categories
Funding
- Army Research Laboratory [W911NF1920209]
- Air Force Office of Scientific Research [21USCOR004, FA9550-18-1-0455]
Ask authors/readers for more resources
This paper proposes six requirements for future measures of trust assessment in automation, and emphasizes the importance of leveraging existing research and considering the needs of practitioners in order to realize the benefits of trust assessment. It also highlights the challenges of trust assessment outside the laboratory environment.
Trust in automation is a foundational principle in Human Factors Engineering. An understanding of trust can help predict and alter much of human-machine interaction (HMI). However, despite the utility of assessing trust in automation in applied settings, there are inherent and unique challenges in trust assessment for those who seek to do so outside of the confines of the sterile lab environment. Because of these challenges, new approaches for trust in automation assessment need to be developed to best suit the unique demands of trust assessment in the real world. This paper lays out six requirements for these future measures: they should (1) be short, unobtrusive, and interaction-based, (2) be context-specific and adaptable, (3) be dynamic, (4) account for autonomy versus automation dependency, (5) account for task dependency, and (6) account for levels of risk. For the benefits of trust assessment to be realized in the real world, future research needs to leverage the existing body of literature on trust in automation while looking toward the needs of the practitioner.
Authors
I am an author on this paper
Click your name to claim this paper and add it to your profile.
Reviews
Recommended
No Data Available