Cover Image for Maximizing the value of reinforcement learning
Cover Image for Maximizing the value of reinforcement learning
Avatar for South Park Commons
Presented by
South Park Commons
South Park Commons helps you get from -1 to 0. To learn more or apply, visit southparkcommons.com.
16 Going

Maximizing the value of reinforcement learning

Register to See Address
San Francisco, California
Registration
Approval Required
Your registration is subject to approval by the host.
Welcome! To join the event, please register below.
About Event

Despite decades of research in RL on sample efficiency and generalization, the scaling of RL currently centers around policy gradient methods that often eschew not only these prior innovations, but even value functions. What aspects of the pre-LLM RL field are most promising to revisit? Is there a place for temporal difference based methods in the transformer-based era of RL?

Join Vmax and SPC for a panel discussion featuring Danijar Hafner (creator of Dreamer), Ashvin Nair (ex-OpenAI, now RL foundations at Cursor) and Nate Rahn (research scientist at Anthropic) on which parts of pre-LLM reinforcement learning are likely to be the most fruitful in the LLM era.

Schedule:

5:00pm doors open
5:30pm - 6:30pm panel discussion
6:30pm - 7:30pm drinks

Who should attend: Researchers and engineers interested in RL.

Location
Please register to see the exact location of this event.
San Francisco, California
Avatar for South Park Commons
Presented by
South Park Commons
South Park Commons helps you get from -1 to 0. To learn more or apply, visit southparkcommons.com.
16 Going