Trustworthy ai

Securing & Governing Autonomous AI Agents: Risks & Safeguards

Securing & Governing Autonomous AI Agents: Risks & Safeguards

Experts Jeff Crume and Josh Spurgin explore the critical security and governance challenges posed by autonomous AI agents. They detail common threats like prompt injection, data poisoning, and model theft, and discuss governance issues such as bias, transparency, and accountability, providing a set of actionable safeguards to build secure, trustworthy, and compliant AI systems.

915: How to Jailbreak LLMs (and How to Prevent It) — with Michelle Yi

915: How to Jailbreak LLMs (and How to Prevent It) — with Michelle Yi

Tech leader and investor Michelle Yi discusses the critical technical aspects of building trustworthy AI systems. She delves into adversarial attack and defense mechanisms, including red teaming, data poisoning, prompt stealing, and "slop squatting," and explores how advanced concepts like Constitutional AI and World Models can create safer, more reliable AI.