We found a match
Your institution may have access to this item. Find your institution then sign in to continue.
- Title
Learning state importance for preference-based reinforcement learning.
- Authors
Zhang, Guoxi; Kashima, Hisashi
- Abstract
Preference-based reinforcement learning (PbRL) develops agents using human preferences. Due to its empirical success, it has prospect of benefiting human-centered applications. Meanwhile, previous work on PbRL overlooks interpretability, which is an indispensable element of ethical artificial intelligence (AI). While prior art for explainable AI offers some machinery, there lacks an approach to select samples to construct explanations. This becomes an issue for PbRL, as transitions relevant to task solving are often outnumbered by irrelevant ones. Thus, ad-hoc sample selection undermines the credibility of explanations. The present study proposes a framework for learning reward functions and state importance from preferences simultaneously. It offers a systematic approach for selecting samples when constructing explanations. Moreover, the present study proposes a perturbation analysis to evaluate the learned state importance quantitatively. Through experiments on discrete and continuous control tasks, the present study demonstrates the proposed framework's efficacy for providing interpretability without sacrificing task performance.
- Subjects
REINFORCEMENT learning; REWARD (Psychology); ARTIFICIAL intelligence; TASK performance
- Publication
Machine Learning, 2024, Vol 113, Issue 4, p1885
- ISSN
0885-6125
- Publication type
Article
- DOI
10.1007/s10994-022-06295-5