Policy Distillation
Introduction
[Paper]: Policy Distillation
The following statements from the paper are key to understand this technique:
- Distillation is a method to transfer knowledge from a teacher model $T$ to a student model $S$.
- Goals:
- It is “used to extract the policy of a reinforcement learning agent and train a new network that performs at the expert level while being dramatically smaller and more efficient.”
- It is “used to consolidate multiple task-specific policies into a single policy.”
The following part has not been finished yet.
Single-Game Policy Distillation
Multi-Task Policy Distillation
This post is licensed under CC BY 4.0 by the author.