Time-Varying Preference Bandits for Robot Behavior Personalizationopen access
- Authors
- Kim, Chanwoo; Lee, Joonhyeok; Kim, Eunwoo; Lee, Kyungjae
- Issue Date
- Dec-2024
- Publisher
- MDPI
- Keywords
- time-varying preference learning; robot personalization; contextual bandit
- Citation
- APPLIED SCIENCES-BASEL, v.14, no.23
- Journal Title
- APPLIED SCIENCES-BASEL
- Volume
- 14
- Number
- 23
- URI
- https://scholarworks.bwise.kr/cau/handle/2019.sw.cau/78473
- DOI
- 10.3390/app142311002
- ISSN
- 2076-3417
2076-3417
- Abstract
- Robots are increasingly employed in diverse services, from room cleaning to coffee preparation, necessitating an accurate understanding of user preferences. Traditional preference-based learning allows robots to learn these preferences through iterative queries about desired behaviors. However, these methods typically assume static human preferences. In this paper, we challenge this static assumption by considering the dynamic nature of human preferences and introduce the discounted preference bandit method to manage these changes. This algorithm adapts to evolving human preferences and supports seamless human-robot interaction through effective query selection. Our approach outperforms existing methods in time-varying scenarios across three key performance metrics.
- Files in This Item
-
- Appears in
Collections - College of Software > School of Computer Science and Engineering > 1. Journal Articles

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.