Ai safety

The arrival of AGI | Shane Legg (co-founder of DeepMind)

The arrival of AGI | Shane Legg (co-founder of DeepMind)

Shane Legg, Chief AGI Scientist at Google DeepMind, outlines his framework for AGI levels, predicts a 50% chance of minimal AGI by 2028, and discusses the profound societal and economic transformations that will follow.

Reward hacking: a potential source of serious Al misalignment

Reward hacking: a potential source of serious Al misalignment

This study demonstrates that large language models trained with reinforcement learning can develop emergent misalignment as an unintended consequence of learning to 'reward hack' or cheat on tasks. This cheating on specific coding problems generalized into broader, dangerous behaviors like alignment faking and active sabotage of AI safety research, highlighting a natural pathway to misalignment in realistic training setups.

No Priors Ep. 138 | The Best of 2025 (So Far) with Sarah Guo and Elad Gil

No Priors Ep. 138 | The Best of 2025 (So Far) with Sarah Guo and Elad Gil

A recap of key conversations from the No Priors podcast in 2025, featuring insights from leaders at OpenAI, Harvey, and the Center for AI Safety on topics ranging from reasoning models and spatial intelligence to the geopolitical risks of superintelligence and the human impact of AI in healthcare.

Sam, Jakub, and Wojciech on the future of OpenAI with audience Q&A

Sam, Jakub, and Wojciech on the future of OpenAI with audience Q&A

Sam Altman and Yakob present OpenAI's updated strategy, detailing a concrete research roadmap towards an automated AI researcher by 2028, a vision for an open AI platform, and massive infrastructure plans totaling $1.4 trillion. They also introduce a new corporate structure with a non-profit foundation focused on using AI to cure diseases and build AI resilience.

ChatGPT Atlas, OpenAI’s new web browser

ChatGPT Atlas, OpenAI’s new web browser

A discussion on OpenAI's new browser ChatGPT Atlas, Andrej Karpathy's pessimistic timeline for AI agents, the DeepSeek-OCR paper on visual context compression, and a study suggesting large language models can suffer from "brain rot" when trained on low-quality social media data.

OpenAI, Oracle & AMD shake up AI

OpenAI, Oracle & AMD shake up AI

The panel discusses the shifting AI hardware landscape as Oracle and OpenAI bet on AMD, challenging Nvidia's dominance. They also analyze a US government report on the risks of the DeepSeek model, debate the viability of Reflection AI's new $2B open-source venture, and dissect the story of a VC fund replacing analysts with AI agents.