Llm

Rethinking Notebooks Powered by AI

Rethinking Notebooks Powered by AI

Vincent Warmerdam from marimo discusses the recent acquisition by Weights & Biases and the future of Python notebooks. He argues that notebooks should evolve from static scratchpads into dynamic, AI-powered applications, highlighting marimo's features for LLM integration, agentic workflows, and creating interactive, reproducible development environments.

Simple AI Upsells 30% Better Than Trained Reps

Simple AI Upsells 30% Better Than Trained Reps

Founders of Simple AI, Catheryn Li & Zach Kamran, discuss their journey from building consumer apps to creating an AI sales agent that handles inbound calls for major brands. They cover their pivot, the technical challenges of integrating with legacy systems, and how their AI outperforms human reps by leveraging hyper-personalization and rapid A/B testing.

Serverless Panel • N. Coult, R. Kohler, D. Anderson, J. Agarwal, A. Laxmi & J. Dongre

Serverless Panel • N. Coult, R. Kohler, D. Anderson, J. Agarwal, A. Laxmi & J. Dongre

A panel of experts from AWS, G-P, and AntStack discuss the practical impact of Generative AI on software development. They explore how AI is used as an accelerant for productivity, the challenges of applying it to large-scale system design, its role in modernization, and the future implications for developer careers and safety-critical systems.

Tool Calling

Tool Calling

A panel discussion with experts from Arcade, Prosus Group, and MeaningStack who argue that most teams are building agents incorrectly. They dissect the failures of simple API wrappers, the pros and cons of MCP, and the critical role of governance and organizational structure in deploying agents successfully.

What is Multimodal RAG? Unlocking LLMs with Vector Databases

What is Multimodal RAG? Unlocking LLMs with Vector Databases

A technical breakdown of three distinct approaches for implementing Multimodal Retrieval-Augmented Generation (RAG), moving from simple text conversion to fully integrated cross-modal systems. The discussion covers the architecture, trade-offs, and capabilities of each method.

Inference at Scale:Breaking the Memory Wall

Inference at Scale:Breaking the Memory Wall

Sid Sheth, CEO of d-matrix, details their memory-centric approach to AI inference hardware, focusing on their Digital In-Memory Compute (DIMC) architecture. He explains how DIMC, an augmented SRAM technology, minimizes data movement to solve the memory bottleneck, delivering significant gains in latency and energy efficiency, particularly for the 'decode' phase of large language models.