Ai safety

Threat Intelligence: How Anthropic stops AI cybercrime

Threat Intelligence: How Anthropic stops AI cybercrime

Anthropic's Threat Intelligence team discusses their new report on how AI models are being used in sophisticated cybercrime operations. They cover the concept of "vibe hacking," a large-scale employment scam run by North Korea, and Anthropic’s multi-layered strategy to detect and counteract these threats.

Gen AI pilots fail, GPT-5's hidden prompt revealed, reasoning model flaws and Claude closing chats

Gen AI pilots fail, GPT-5's hidden prompt revealed, reasoning model flaws and Claude closing chats

A deep dive into why most enterprise GenAI pilots are failing, the debate around hidden system prompts in models like GPT-5, new research questioning the reliability of "chain of thought" reasoning, and the controversy over Anthropic's "AI welfare" justification for shutting down conversations.

Anthropic Co-founder: Building Claude Code, Lessons From GPT-3 & LLM System Design

Anthropic Co-founder: Building Claude Code, Lessons From GPT-3 & LLM System Design

Tom Brown, co-founder of Anthropic, shares his journey from a YC founder to a key figure behind AI's scaling breakthroughs. He discusses the discovery of scaling laws that underpinned GPT-3, the mission-driven founding of Anthropic, the surprising success of Claude for coding, and his perspective on what he calls "humanity's largest infrastructure buildout ever."

Interpretability: Understanding how AI models think

Interpretability: Understanding how AI models think

Members of Anthropic's interpretability team discuss their research into the inner workings of large language models. They explore the analogy of studying AI as a biological system, the surprising discovery of internal "features" or concepts, and why this research is critical for understanding model behavior like hallucinations, sycophancy, and long-term planning, ultimately aiming to ensure AI safety.

The Current Reality of American AI Policy: From ‘Pause AI’ to ‘Build’

The Current Reality of American AI Policy: From ‘Pause AI’ to ‘Build’

a16z's Martin Casado and Anjney Midha detail the dramatic shift in U.S. AI policy from a 'pause AI' stance, fueled by doomerism and flawed analogies, to a pro-innovation 'win the race' strategy. They discuss how China's progress shattered illusions of a U.S. lead, the strategic business case for open source AI, and the pragmatic promise of the new AI Action Plan.

Dwarkesh and Noah Smith on AGI and the Economy

Dwarkesh and Noah Smith on AGI and the Economy

Dwarkesh Patel and Noah Smith debate the definition of AGI, its economic implications, and timelines. They contrast an economic definition (automating white-collar work) with a cognitive one, exploring why current models lack economic value despite reasoning abilities due to a failure in 'continual learning'. The discussion covers the potential for explosive economic growth versus a collapse in consumer demand, the substitution vs. complementarity of human labor, and the geopolitical shift from population size to inference capacity as the basis of power.