Cover Image for OpenClaw-RL: The Self-Evolving Agent (NICE No.158)
Cover Image for OpenClaw-RL: The Self-Evolving Agent (NICE No.158)
Hosted By
6 Went

OpenClaw-RL: The Self-Evolving Agent (NICE No.158)

Hosted by NICE AI Talk
YouTube
Registration
Past Event
Welcome! To join the event, please register below.
About Event

Welcome to NICE TALK 158! | OpenClaw-RL: The Self-Evolving Agent

What if your model could learn and evolve from every interaction after deployment?

This talk introduces OpenClaw-RL, a novel reinforcement learning framework that enables agents to self-improve autonomously during real-world use. OpenClaw-RL functions as an RL server—users simply deploy their personal models on it, and the models automatically and continuously optimize throughout usage. We propose an optimization method that combines the strengths of GRPO and On-policy Distillation, transforming the entire history of model-user-environment interactions into RL training signals. The result: personal agents that don't just stay static, but grow more capable and adaptive the more they are used. We also design interesting experiments to validate the framework's efficient optimization capability for personal agents.

Resources:

Speaker: Yinjie Wang

Yinjie Wang is a second-year PhD student at the University of Chicago and a Research Intern at the Princeton AI Lab. Previously, he graduated from the Special Class for the Gifted Young (SCGY) at the University of Science and Technology of China. His research focuses on LLMs, autonomous agents, and reinforcement learning. He is the creator of several open-source RL frameworks including OpenClaw-RL, RLAnything, CURE, and dLLM-RL. His work has been published at NeurIPS, ICLR, and other top-tier venues, receiving a Spotlight award at NeurIPS 2025.

Personal website: https://yinjjiew.github.io/

Host: Wenyue Hua

Senior Researcher at Microsoft Research.

Hosted By
6 Went