#prompt-injection
7 episodes
#2691: Can AI Agents Safely Manage Your API Keys?
Is it time to let AI agents handle your API key creation and rotation? We explore the real security tradeoffs.
#2472: AI Gateways: Where Guardrails Actually Break
PII detection at the gateway layer can block legitimate invoices. Here's how guardrails actually work and where they fail.
#2180: The Sandboxing Tradeoff in Agent Design
AI agents need broad permissions to be useful—but every permission expands the attack surface. We map the real threat landscape and the isolation t...
#1957: Why AI Agents Think in Circles, Not Lines
Linear AI pipelines are brittle. Learn why loops, reflection, and state management are the new standard for reliable, autonomous agents.
#1217: Stop the Leak: Securing Your AI’s System Instructions
Discover why AI models leak their secret instructions and how to defend your intellectual property using modern prompt hardening techniques.
#1070: The Agentic Secret Gap: Securing the AI Developer Workflow
AI agents write code in seconds, but manual secret management is a major bottleneck. Explore how to bridge the gap between speed and security.
#44: AI's Wild West: Battling Injection & Poisoning
AI's Wild West: Battling prompt injection and poisoning. Discover how AI threats are shifting from sci-fi to insidious attacks on the models...