Inverse reinforcement learning through max-margin algorithm

Syed Ihtesham Hussain Shah, Antonio Coronato

Research output: Chapter in Book / Report / Conference proceedingChapterAcademicpeer-review

Abstract

Reinforcement Learning (RL) methods provide a solution for decisionmaking problems under uncertainty. An agent finds a suitable policy through a reward function by interacting with a dynamic environment. However, for complex and large problems it is very difficult to specify and tune the reward function. Inverse Reinforcement Learning (IRL) may mitigate this problem by learning the reward function through expert demonstrations. This work exploits an IRL method named Max-Margin Algorithm (MMA) to learn the reward function for a robotic navigation problem. The learned reward function reveals the demonstrated policy (expert policy) better than all other policies. Results show that this method has better convergence and learned reward functions through the adopted method represents expert behavior more efficiently.
Original languageEnglish
Title of host publicationIntelligent Environments 2021
Subtitle of host publicationWorkshop Proceedings of the 17th International Conference on Intelligent Environments
PublisherIOS Press
Pages190-201
Number of pages12
ISBN (Electronic)9781643681870
ISBN (Print)9781643681863
DOIs
Publication statusPublished - 2021
Externally publishedYes

Publication series

NameAmbient Intelligence and Smart Environments
PublisherIOS Press
Volume29

Fingerprint

Dive into the research topics of 'Inverse reinforcement learning through max-margin algorithm'. Together they form a unique fingerprint.

Cite this