Abstract

While Reinforcement Learning ( RL) has made great strides towards solving increasinglycomplicated problems, many algorithms are still brittle to even slight environmental changes.Contextual Reinforcement Learning (cRL) provides a framework to model such changes ina principled manner, thereby enabling flexible, precise and interpretable task specificationand generation. Our goal is to show how the framework of cRL contributes to improvingzero-shot generalization in RL through meaningful benchmarks and structured reasoningabout generalization tasks. We confirm the insight that optimal behavior in cRL requirescontext information, as in other related areas of partial observability. To empirically validatethis in the cRL framework, we provide various context-extended versions of common RLenvironments. They are part of the first benchmark library, CARL, designed for generalizationbased on cRL extensions of popular benchmarks, which we propose as a testbed to furtherstudy general agents. We show that in the contextual setting, even simple RL environmentsbecome challenging - and that naive solutions are not enough to generalize across complexcontext spaces.

Links and resources

Tags

community

  • @tntl3s
  • @ail3s
@ail3s's tags highlighted