TU Darmstadt / ULB / TUprints

Inverse Reinforcement Learning via Nonparametric Spatio-Temporal Subgoal Modeling

Šošić, Adrian ; Rueckert, Elmar ; Peters, Jan ; Zoubir, Abdelhak M. ; Koeppl, Heinz (2024)
Inverse Reinforcement Learning via Nonparametric Spatio-Temporal Subgoal Modeling.
In: Journal of Machine Learning Research, 2018, 19 (69)
doi: 10.26083/tuprints-00026700
Article, Secondary publication, Publisher's Version

[img] Text
18-113.pdf
Copyright Information: CC BY 4.0 International - Creative Commons, Attribution.

Download (2MB)
Item Type: Article
Type of entry: Secondary publication
Title: Inverse Reinforcement Learning via Nonparametric Spatio-Temporal Subgoal Modeling
Language: English
Date: 30 April 2024
Place of Publication: Darmstadt
Year of primary publication: 2018
Place of primary publication: Brookline, Massachusetts
Publisher: Microtome Publishing
Journal or Publication Title: Journal of Machine Learning Research
Volume of the journal: 19
Issue Number: 69
Collation: 45 Seiten
DOI: 10.26083/tuprints-00026700
Corresponding Links:
Origin: Secondary publication service
Abstract:

Advances in the field of inverse reinforcement learning (IRL) have led to sophisticated inference frameworks that relax the original modeling assumption of observing an agent behavior that reflects only a single intention. Instead of learning a global behavioral model, recent IRL methods divide the demonstration data into parts, to account for the fact that different trajectories may correspond to different intentions, e.g., because they were generated by different domain experts. In this work, we go one step further: using the intuitive concept of subgoals, we build upon the premise that even a single trajectory can be explained more efficiently locally within a certain context than globally, enabling a more compact representation of the observed behavior. Based on this assumption, we build an implicit intentional model of the agent's goals to forecast its behavior in unobserved situations. The result is an integrated Bayesian prediction framework that significantly outperforms existing IRL solutions and provides smooth policy estimates consistent with the expert's plan. Most notably, our framework naturally handles situations where the intentions of the agent change over time and classical IRL algorithms fail. In addition, due to its probabilistic nature, the model can be straightforwardly applied in active learning scenarios to guide the demonstration process of the expert.

Uncontrolled Keywords: Learning from Demonstration, Inverse Reinforcement Learning, Bayesian Nonparametric Modeling, Subgoal Inference, Graphical Models, Gibbs Sampling
Status: Publisher's Version
URN: urn:nbn:de:tuda-tuprints-267009
Classification DDC: 000 Generalities, computers, information > 004 Computer science
500 Science and mathematics > 570 Life sciences, biology
600 Technology, medicine, applied sciences > 621.3 Electrical engineering, electronics
Divisions: 18 Department of Electrical Engineering and Information Technology > Institute for Telecommunications > Bioinspired Communication Systems
18 Department of Electrical Engineering and Information Technology > Self-Organizing Systems Lab
18 Department of Electrical Engineering and Information Technology > Institute for Telecommunications > Signal Processing
20 Department of Computer Science > Intelligent Autonomous Systems
Date Deposited: 30 Apr 2024 09:17
Last Modified: 02 Aug 2024 08:21
URI: https://tuprints.ulb.tu-darmstadt.de/id/eprint/26700
PPN: 520266900
Export:
Actions (login required)
View Item View Item