Modeling Variation in Human Feedback with User Inputs: An Exploratory Methodology
To expedite the development process of interactive reinforcement learning (IntRL) algorithms, prior work often uses perfect oracles as simulated human teachers to furnish feedback signals. These oracles typically derive from ground-truth knowledge or optimal policies, providing dense and error-free...
Saved in:
| Published in: | 2024 19th ACM/IEEE International Conference on Human-Robot Interaction (HRI) pp. 303 - 312 |
|---|---|
| Main Authors: | , , |
| Format: | Conference Proceeding |
| Language: | English |
| Published: |
ACM
11.03.2024
|
| Subjects: | |
| Online Access: | Get full text |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Be the first to leave a comment!