Large language models

Scaling and the Road to Human-Level AI | Anthropic Co-founder Jared Kaplan

Scaling and the Road to Human-Level AI | Anthropic Co-founder Jared Kaplan

Jared Kaplan, co-founder of Anthropic, explains how the discovery of predictable, physics-like scaling laws in AI training provides a clear roadmap for progress. He details the two main phases of model training (pre-training and RL), discusses how scaling compute predictably unlocks longer-horizon task capabilities, and outlines the remaining challenges—memory, nuanced oversight, and organizational knowledge—on the path to human-level AI.

Building Better Language Models Through Global Understanding

Building Better Language Models Through Global Understanding

Dr. Mazi Fadai discusses the critical challenges in multilingual AI, including data imbalances and flawed evaluation methodologies. She argues that tackling these difficult multilingual problems is not only essential for global accessibility but also a catalyst for fundamental AI innovation, much like how machine translation research led to the Transformer architecture. The talk introduces new, more culturally aware evaluation benchmarks like Global MMLU and INCLUDE as a path toward building more robust and globally representative language models.

Intelligence = Doing More with Less (David Krakauer)

Intelligence = Doing More with Less (David Krakauer)

Prof. David Krakauer argues that we are confusing knowledge with intelligence. He critiques the AI community's superficial definition of "emergence" in LLMs, contrasting it with the true meaning from complex systems: a fundamental change in internal organization that allows for a simpler, more powerful macroscopic description. He introduces "exbodiment"—outsourcing cognition to external tools—as a key part of collective intelligence, but warns that our evolutionary drive to conserve energy will lead us to outsource our thinking to AI, causing a "diminution and dilution" of human thought.

Inside GPT – The Maths Behind the Magic • Alan Smith • GOTO 2024

Inside GPT – The Maths Behind the Magic • Alan Smith • GOTO 2024

A deep dive into the internal workings of Large Language Models like GPT, explaining the journey from a text prompt through tokenization, embeddings, and the attention mechanism to generate a response.

OpenAI Just Released ChatGPT Agent, Its Most Powerful Agent Yet

OpenAI Just Released ChatGPT Agent, Its Most Powerful Agent Yet

The OpenAI team details the creation of a new, powerful AI agent in ChatGPT, achieved by unifying the Deep Research and Operator models. They cover its unified architecture with shared state across tools, the reinforcement learning techniques used for training, and the critical safety measures required for an agent that can take real-world actions.

The Future of Software Development - Vibe Coding, Prompt Engineering & AI Assistants

The Future of Software Development - Vibe Coding, Prompt Engineering & AI Assistants

A discussion on the state of technical infrastructure, focusing on how AI and Large Language Models represent a new, fourth foundational pillar alongside compute, network, and storage. The talk covers how AI is disrupting software itself, the investment landscape, and the future of the developer profession.