Learning Hierarchical Policies in Dynamic Environments
Advisor: Ruslan Salakhutdinov, School of Computer Science, Carnegie Mellon University
- Proposed a hierarchical RL and meta RL based framework for solving sparse rewards tasks in dynamic environments.
- The agent first learns a generic representation of a set of skills over a distribution of environments using meta learning. These skills are then fine-tuned to the given environment with a few gradient updates and a high level policy over these skills is learned for solving the required task.
The project report is available here.