Finding scaling laws for Reinforcement Learning

preview_player
Показать описание
Neural scaling laws have received a lot of attention in recent ML research, ever since it was discovered that generative language models improve their performance as a power law of available resources. Since then, power-law scaling laws seem to pop up in every field and setting imaginable. These laws provide clear instructions on how to train large models on multi-million dollar budgets, and have directly guided the creation of SOTA models like GPT-3. Despite all this, Reinforcement Learning had until recently almost no record of power-law scaling.

In this talk, Oren Neumann shares how his team found power-law scaling laws for AlphaZero, why any previous attempt to find these laws in RL failed, and how to train a model efficiently when a single training attempt costs several million dollars.

Bio: Oren Neumann is PhD candidate in Complex Systems physics at Goethe University in Frankfurt. Coming from a background in Condensed Matter physics, where emergent power laws have been studied for decades, he focused his research on scaling phenomena in Reinforcement Learning.

Рекомендации по теме