Designing AI agents to resist prompt injection
OpenAI outlines the architectural patterns used to defend ChatGPT agents against prompt injection and social engineering — constraining risky actions, separating trusted from untrusted content, and protecting sensitive data in agent workflows. Directly applicable to any developer building agents that process external content.





