Effects of a Social Robot's Self-Explanations on How Humans Understand and Evaluate Its Behavior
Social robots interacting with users in real-life environments will often show surprising or even undesirable behavior. In this paper we investigate whether a robot's ability to self-explain its behavior affects the users' perception and assessment of this behavior. We propose an explanati...
Uložené v:
| Vydané v: | 2020 15th ACM/IEEE International Conference on Human-Robot Interaction (HRI) s. 619 - 627 |
|---|---|
| Hlavní autori: | , |
| Médium: | Konferenčný príspevok.. |
| Jazyk: | English |
| Vydavateľské údaje: |
New York, NY, USA
ACM
09.03.2020
|
| Edícia: | ACM Conferences |
| Predmet: | |
| ISBN: | 1450367461, 9781450367462 |
| ISSN: | 2167-2148 |
| On-line prístup: | Získať plný text |
| Tagy: |
Pridať tag
Žiadne tagy, Buďte prvý, kto otaguje tento záznam!
|
| Shrnutí: | Social robots interacting with users in real-life environments will often show surprising or even undesirable behavior. In this paper we investigate whether a robot's ability to self-explain its behavior affects the users' perception and assessment of this behavior. We propose an explanation model based on humans' folk-psychological concepts and test different explanation strategies in specifically designed HRI scenarios with robot behaviors perceived as intentional, but differently surprising or desirable. All types of explanation strategies increased the understandability and desirability of the behaviors. While merely stating an action had similar effects as giving a reason for it (an intention or need), combining both in a causal explanation helped the robot to better justify its behavior and to increase its understandability and desirability to a larger extent. |
|---|---|
| ISBN: | 1450367461 9781450367462 |
| ISSN: | 2167-2148 |
| DOI: | 10.1145/3319502.3374802 |

