Gpu

931: Boost Your Profits with Mathematical Optimization, feat. Jerry Yurchisin

931: Boost Your Profits with Mathematical Optimization, feat. Jerry Yurchisin

Gurobi's Jerry Yurchisin explains the power of mathematical optimization, a prescriptive approach that complements AI's predictive capabilities. This summary covers how to get started with free resources, the use of GPUs and LLMs to enhance optimization, real-world applications at companies like Toyota, and its relationship with quantum computing.

Why Your Cloud Isn't Ready for Production AI

Why Your Cloud Isn't Ready for Production AI

Zhen Lu, CEO of Runpod, discusses the shift from Web 2.0 architectures to an "AI-first" cloud. The conversation covers the unique hardware and software requirements for production AI, key use cases like generative media and enterprise agents, and the critical challenges of reliability and operationalization in the new AI stack.

How To Train An LLM with Anthropic's Head of Pretraining

How To Train An LLM with Anthropic's Head of Pretraining

Anthropic's Head of Pre-training, Nick Joseph, details the immense engineering and infrastructure challenges behind training frontier models like Claude. He covers the evolution from early-stage custom frameworks to debugging hardware at massive scale, balancing pre-training with RL, and the strategic importance of data quality and team composition.

921: NPUs vs GPUs vs CPUs for Local AI Workloads — with Dell’s Ish Shah and Shirish Gupta

921: NPUs vs GPUs vs CPUs for Local AI Workloads — with Dell’s Ish Shah and Shirish Gupta

Shirish Gupta and Ish Shah from Dell Technologies explore the evolving landscape of AI hardware. They discuss why Windows, enhanced by WSL 2, remains a dominant platform for developers, and delve into the distinct roles of CPUs, GPUs, and the increasingly important Neural Processing Units (NPUs). The conversation covers the trade-offs between local and cloud computing for AI workloads and introduces new hardware, like workstations with discrete NPUs, that are making on-device AI more powerful and accessible than ever.

Monster prompt, OpenAI’s business play, nano-banana and US Open experimentations

Monster prompt, OpenAI’s business play, nano-banana and US Open experimentations

The panel discusses KPMG's 100-page prompt for its TaxBot, debating the future of prompt engineering versus fine-tuning. They also analyze OpenAI's potential move into selling cloud infrastructure, the impressive capabilities of Google's new image model, Nano-Banana, and new AI-powered fan experiences at the US Open.

The Truth About LLM Training

The Truth About LLM Training

Paul van der Boor and Zulkuf Genc from Prosus discuss the practical realities of deploying AI agents in production. They cover their in-house evaluation framework, strategies for navigating the GPU market, the importance of fine-tuning over building from scratch, and how they use AI to analyze usage patterns in a privacy-preserving manner.