Training Agents with Reinforcement Learning: Kyle Corbitt
Автор: Open Data Science and AI Conference
Загружено: 2025-12-17
Просмотров: 80
In this episode, we speak with Kyle Corbitt, co-founder and CEO of OpenPip, recently acquired by CoreWeave, to explore the evolving role of reinforcement learning in building smarter, more reliable AI agents. Kyle shares the journey of OpenPipe from supervised fine-tuning to developing ART (Agent Reinforcement Trainer), their open-source RL toolkit designed to train AI agents that can think, adapt, and perform with greater autonomy. The discussion spans technical insights, practical applications, startup lessons from YC’s Startup School, and the future of agent-based AI systems.
Key Topics Covered:
Why reinforcement learning is gaining attention in modern Agent development
The transition from supervised fine-tuning (SFT) to reinforcement learning (RL)
Practical differences between RL and SFT, including weight movement and model reliability
OpenPipe’s approach with ART: supporting multi-turn agent training and tool use
How ART differs from OpenAI’s RFT implementation
The importance of consistent agent behavior in production and how RL helps
Avoiding reward hacking and the role of Ruler, OpenPipe’s LLM-based judging system
Cost-efficiency strategies in RL training using serverless infrastructure
OpenPipe’s long-term vision for self-improving agents
Advice for AI startup founders on building in a rapidly evolving ecosystem
Visit our website and choose the nearest ODSC event to attend and experience all our training and workshops: https://odsc.ai
To watch more videos like this, visit https://aiplus.training
Sign up for the newsletter to stay up to date with the latest trends in data science: https://opendatascience.com/newsletter/
Follow us online!
• Facebook: / opendatasci
• Instagram: / odsc
• Blog: https://opendatascience.com/
• LinkedIn: / open-data-science
• X (twitter): https://x.com/_odsc
Доступные форматы для скачивания:
Скачать видео mp4
-
Информация по загрузке: