Llm

Efficient Reinforcement Learning – Rhythm Garg & Linden Li, Applied Compute

Efficient Reinforcement Learning – Rhythm Garg & Linden Li, Applied Compute

A deep dive into the challenges and solutions for efficient Reinforcement Learning (RL) in enterprise settings. The talk contrasts synchronous and asynchronous RL, explains the critical trade-off of "staleness" versus stability, and details a first-principles system model used to optimize GPU allocation for maximum throughput.

Compilers in the Age of LLMs — Yusuf Olokoba, Muna

Compilers in the Age of LLMs — Yusuf Olokoba, Muna

Yusuf Olokoba, founder of Muna, details a compiler-based approach to transform Python AI functions into self-contained native binaries. This talk explores the technical pipeline, including custom AST-based tracing, type propagation, and the strategic use of LLMs for code generation, enabling a universal, OpenAI-style client for running any model on any platform.

The Unbearable Lightness of Agent Optimization — Alberto Romero, Jointly

The Unbearable Lightness of Agent Optimization — Alberto Romero, Jointly

This talk introduces Meta-ACE, a learned meta-optimization framework that dynamically orchestrates multiple strategies (context evolution, adaptive compute, hierarchical verification, and more) to maximize AI agent performance. The framework profiles each task to select an optimal strategy bundle, overcoming the single-dimension limitations of previous methods.

Inside the AI Black Box

Inside the AI Black Box

Emmanuel Ameisen of Anthropic's interpretability team explains the inner workings of LLMs, drawing analogies to biology. He covers surprising findings on how models plan, represent concepts across languages, and the mechanistic causes of hallucinations, offering practical advice for developers on evaluation and post-training strategies.

I’m Teaching AI Self-Improvement Techniques

I’m Teaching AI Self-Improvement Techniques

Aman Khan from Arize discusses the challenges of building reliable AI agents and introduces a novel technique called "metaprompting". This method uses continuous, natural language feedback to optimize an agent's system prompt, effectively training its "memory" or context, leading to significant performance gains even for smaller models.

AI ATTACKS! How Hackers Weaponize Artificial Intelligence

AI ATTACKS! How Hackers Weaponize Artificial Intelligence

AI is no longer just a tool for defense; it's being weaponized by malicious actors. This summary explores six emerging AI-powered cyber attacks, from automated login attempts and polymorphic ransomware to hyper-personalized phishing and deepfake fraud. It details how AI agents and Large Language Models (LLMs) are used to automate the entire attack kill chain, significantly lowering the skill barrier for attackers and necessitating an evolution in cyber defense strategies.