Illia Polosukhin Urges Stronger Oversight for AI Agents
Illia Polosukhin, co-author of the original transformer paper and co-founder of Near Protocol, warns that as AI agents gain autonomy and capability, existing institutions are not prepared to manage their systemic risks. He calls for clearer accountability, human-in-the-loop controls, and institutional readiness to govern agent behaviors, failure modes, and economic impacts. For practitioners, this reframes agent development as not only an engineering problem but also an operational and governance challenge: teams must design verification, auditability, and escalation paths into agent workflows before deployment at scale.
What happened
Illia Polosukhin, co-author of the seminal transformer paper and co-founder of Near Protocol, warned that rising capability in AI agents requires stronger institutional oversight and operational preparation. He argues that autonomous agents will expose gaps in governance, accountability, and verification if organizations treat agent deployment purely as a product engineering task.
Technical details
Practitioners should focus on concrete controls that make agent behavior observable and modifiable. Key measures include:
- •Designing human-in-the-loop checkpoints for high-risk decisions, with clear escalation and intervention paths.
- •Building continuous audit trails and tamper-evident logs to support post-hoc analysis and incident forensics.
- •Implementing verification harnesses and sandboxed testing that exercise economic, safety, and specification-edge cases before live rollout.
Context and significance
The call matters because agents shift risk from isolated model outputs to persistent, stateful systems that can act across services and markets. Unlike single-shot LLM responses, agents combine planning, execution, and integrated tooling access, increasing potential for cascading failures and unintended economic effects. The shift elevates non-ML concerns-identity, authorization, liability, and operational monitoring-into first-class engineering constraints. Illia Polosukhin's profile as a transformer co-author gives weight to the message within the ML community; the industry is moving from model capability milestones to deployability and governance maturity.
What to watch
Teams should prioritize instrumenting agents for observability and human control, and organizations should align legal, security, and product teams on escalation procedures. Regulators and standards bodies will likely focus next on certification of critical-agent workflows and minimum auditability requirements.
Practitioner takeaway
Treat agents as distributed socio-technical systems, not just model endpoints; bake in auditability, test harnesses, and human oversight before scaling deployments.
Scoring Rationale
The warning comes from a high-profile technical founder and reframes a growing operational challenge for ML teams, but it is cautionary rather than introducing new technology or regulation. The topic is immediately relevant to practitioners deploying agents, hence a notable impact score.
Practice interview problems based on real data
1,500+ SQL & Python problems across 15 industry datasets — the exact type of data you work with.
Try 250 free problemsStep-by-step roadmaps from zero to job-ready — curated courses, salary data, and the exact learning order that gets you hired.



