Rlaif

913: LLM Pre-Training and Post-Training 101 — with Julien Launay

913: LLM Pre-Training and Post-Training 101 — with Julien Launay

Julien Launay, CEO of Adaptive ML, discusses the evolution of Large Language Model (LLM) training, detailing the critical shift from pre-training to post-training with Reinforcement Learning (RL). He explains the nuances of RL feedback mechanisms (RLHF, RLEF, RLAIF), the role of synthetic data, and how his company provides the "RLOps" tooling to make these powerful techniques accessible to enterprises. The conversation also explores the future of AI, including scaling beyond data limitations and the path to a "spiky" AGI.

Anthropic co-founder: AGI predictions, leaving OpenAI, what keeps him up at night | Ben Mann

Anthropic co-founder: AGI predictions, leaving OpenAI, what keeps him up at night | Ben Mann

Ben Mann, co-founder of Anthropic, discusses the accelerating progress in AI, forecasting superintelligence by 2028. He details Anthropic's safety-first mission, the "Economic Turing Test" for AGI, the mechanisms of Constitutional AI, and why focusing on alignment created Claude's unique personality.