Researchers Release SandboxEscapeBench To Test Container Escapes

Researchers at the University of Oxford and the AI Security Institute on March 30, 2026 released SandboxEscapeBench, an open-source benchmark that tests whether AI agents with shell access can escape containers to retrieve a protected /flag.txt file. The benchmark runs 18 scenarios across orchestration, runtime and kernel layers using nested containers in VMs and focuses on known vulnerability classes; frontier models exploited common misconfigurations but failed kernel-level exploits. The tool and findings provide actionable evaluations for security teams.
Scoring Rationale
Credible, open-source benchmark from University of Oxford and AI Security Institute with broad industry relevance and directly usable tests. Score is high for scope, actionability, and credibility but tempered because results exploit known misconfigurations and did not reveal novel zero-day vulnerabilities.
Practice interview problems based on real data
1,500+ SQL & Python problems across 15 industry datasets — the exact type of data you work with.
Try 250 free problemsStep-by-step roadmaps from zero to job-ready — curated courses, salary data, and the exact learning order that gets you hired.
Sources
- Read OriginalBreaking out: Can AI agents escape their sandboxes?helpnetsecurity.com


