Authors
- Zihan Ding* - Princeton University (zhding[at]mail.ustc.edu.cn)
Abstract
To alleviate the low sample efficiency problem in deep reinforcement learning, imitation learning, or called apprenticeship learning, is one of the potential approaches, which leverages the expert demonstrations in sequential decision making process. In order to provide the readers a comprehensive understanding about how to effectively extract information from the demonstration data, we introduce the most important categories in imitation learning, including behavioral cloning, inverse reinforcement learning, imitation learning from observations, probabilistic methods and other methods. Imitation learning can either be regarded as an initialization or a guidance for training the agent in the scope of reinforcement learning. Combination of imitation learning and reinforcement learning is a promising direction for efficient learning and faster policy optimization in practice.
Keywords: imitation learning, apprenticeship learning, demonstration, reinforcement learning, behavioral cloning, inverse reinforcement learning, generative adversarial networks, sample efficiency
Content
Citation
To cite this book, please use this bibtex entry:
@incollection{deepRL-chapter8-2020,
title={Imitation Learning},
chapter={8},
author={Zihan Ding},
editor={Hao Dong, Zihan Ding, Shanghang Zhang},
booktitle={Deep Reinforcement Learning: Fundamentals, Research, and Applications},
publisher={Springer Nature},
pages={273-306},
note={\url{http://www.deepreinforcementlearningbook.org}},
year={2020}
}
If you find any typos or have suggestions for improving the book, do not hesitate to contact with the corresponding author (name with *).