Gpu

Nvidia CTO Michael Kagan: Scaling Beyond Moore's Law to Million-GPU Clusters

Nvidia CTO Michael Kagan: Scaling Beyond Moore's Law to Million-GPU Clusters

Nvidia CTO Michael Kagan explains how the Mellanox acquisition was key to scaling AI infrastructure from single GPUs to million-GPU data centers. He covers the critical role of networking in system performance, the shift from training to inference workloads, and his vision for AI's future in scientific discovery.

931: Boost Your Profits with Mathematical Optimization, feat. Jerry Yurchisin

931: Boost Your Profits with Mathematical Optimization, feat. Jerry Yurchisin

Gurobi's Jerry Yurchisin explains the power of mathematical optimization, a prescriptive approach that complements AI's predictive capabilities. This summary covers how to get started with free resources, the use of GPUs and LLMs to enhance optimization, real-world applications at companies like Toyota, and its relationship with quantum computing.

Why Your Cloud Isn't Ready for Production AI

Why Your Cloud Isn't Ready for Production AI

Zhen Lu, CEO of Runpod, discusses the shift from Web 2.0 architectures to an "AI-first" cloud. The conversation covers the unique hardware and software requirements for production AI, key use cases like generative media and enterprise agents, and the critical challenges of reliability and operationalization in the new AI stack.

How To Train An LLM with Anthropic's Head of Pretraining

How To Train An LLM with Anthropic's Head of Pretraining

Anthropic's Head of Pre-training, Nick Joseph, details the immense engineering and infrastructure challenges behind training frontier models like Claude. He covers the evolution from early-stage custom frameworks to debugging hardware at massive scale, balancing pre-training with RL, and the strategic importance of data quality and team composition.

921: NPUs vs GPUs vs CPUs for Local AI Workloads — with Dell’s Ish Shah and Shirish Gupta

921: NPUs vs GPUs vs CPUs for Local AI Workloads — with Dell’s Ish Shah and Shirish Gupta

Shirish Gupta and Ish Shah from Dell Technologies explore the evolving landscape of AI hardware. They discuss why Windows, enhanced by WSL 2, remains a dominant platform for developers, and delve into the distinct roles of CPUs, GPUs, and the increasingly important Neural Processing Units (NPUs). The conversation covers the trade-offs between local and cloud computing for AI workloads and introduces new hardware, like workstations with discrete NPUs, that are making on-device AI more powerful and accessible than ever.

Monster prompt, OpenAI’s business play, nano-banana and US Open experimentations

Monster prompt, OpenAI’s business play, nano-banana and US Open experimentations

The panel discusses KPMG's 100-page prompt for its TaxBot, debating the future of prompt engineering versus fine-tuning. They also analyze OpenAI's potential move into selling cloud infrastructure, the impressive capabilities of Google's new image model, Nano-Banana, and new AI-powered fan experiences at the US Open.