Modeling Variation in Human Feedback with User Inputs: An Exploratory Methodology

To expedite the development process of interactive reinforcement learning (IntRL) algorithms, prior work often uses perfect oracles as simulated human teachers to furnish feedback signals. These oracles typically derive from ground-truth knowledge or optimal policies, providing dense and error-free...

Full description

Saved in:
Bibliographic Details
Published in:2024 19th ACM/IEEE International Conference on Human-Robot Interaction (HRI) pp. 303 - 312
Main Authors: Huang, Jindan, Aronson, Reuben M., Short, Elaine Schaertl
Format: Conference Proceeding
Language:English
Published: ACM 11.03.2024
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Be the first to leave a comment!
You must be logged in first