Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                

Efficient Transformers in Reinforcement Learning using Actor-Learner Distillation

This work develops an Actor-Learner Distillation procedure that leverages a continual form of distillation that transfers learning progress from a large capacity learner model to a small capacity actor model in the context of partially-observable environments.

Bookmark
Add
Share
Cite
Sun Apr 04 2021
Citations
25
by Emilio Parisotto, R. Salakhutdinov
CHAT WITH RESEARCH


QUESTIONS & ANSWERS

Log in to generate
TL;DR
AI STORY
ABSTRACT
PAPER
This work develops an Actor-Learner Distillation procedure that leverages a continual form of distillation that transfers learning progress from a large capacity learner model to a small capacity actor model in the context of partially-observable environments.


Research is provided by Semantic Scholar and AI-generated text may at times produce inaccurate results.
Information provided on this site does not constitute legal, financial, medical, or any other professional advice.

DATA LICENSING
Search and article data is provided under CC BY-NC or ODC-BY and via The Semantic Scholar Open Data Platform. Read more at Kinney, Rodney Michael et al. “The Semantic Scholar Open Data Platform.” ArXiv abs/2301.10140 (2023): n. pag.