For my undergraduate thesis, I worked on a hierarchical reinforcement learning (HRL) project. The results of the research was published in NeurIPS 2019, with me serving as a primary author of the paper. Check out our paper!
This work would not be available without the help of Siyuan Li, Mingxue Tang, and Prof. Chongjie Zhang. I spent a significant amount of time in the Machine Intelligence Group to accomplish this work.
Please check out our videos!
Hierarchical Reinforcement Learning (HRL) is a promising approach to solving long-horizon problems with sparse and delayed rewards.
Many existing HRL algorithms either use pre-trained low-level skills that are unadaptable, or require domain-specific information to define low-level rewards. In this paper, we aim to adapt low-level skills to downstream tasks while maintaining the generality of reward design. We propose an HRL framework which sets auxiliary rewards for low-level skill training based on the advantage function of the high-level policy.
This auxiliary reward enables efficient, simultaneous learning of the high-level policy and low-level skills without using task-specific knowledge.
In addition, we also theoretically prove that
optimizing low-level skills with this auxiliary reward will increase the task return for the joint policy.
Experimental results show that our algorithm dramatically outperforms other state-of-the-art HRL methods in Mujoco domains. We also find both low-level and high-level policies trained by our algorithm transferable.