The Hidden Risk in AI Delegation
When Metaphors Replace Mechanisms
2/19/20261 min read


AI delegation is having a moment. But we should be careful with one potentially misleading assumption: that delegating to an AI is “basically like” delegating to a person.
A recent paper on Intelligent AI Delegation by Google DeepMind (see link below) frames delegation with human concepts: roles, responsibility, accountability, trust, reputation, escalation.
I like the direction. At the same time, we should remember what an AI agent actually is: a PROBABILISTIC OPTIMIZER WITH TOOL ACCESS. Meaning such an approach contains risks that should be addressed:
1) “Authority” isn’t social for agents, it’s technical
For humans, authority is enforced by norms and institutions.
For agents, “authority” is permissions, credentials, tool access. If we don’t translate the metaphor into controls (least privilege, revocation, sandboxing), we risk safety-by-storytelling.
2) “Trust” and “reputation” are gameable by default
In an agent ecosystem, identities are cheap, agents are copyable, and “good behavior” can be selectively shown to measurement channels. Trust needs identity + attestation to be operational, not aspirational.
3) Monitoring won’t catch the many small failures
Many failures are silent: plausible-but-wrong outputs, subtle drift, partial tool misuse. Trigger-based “detect → diagnose → re-delegate” can overestimate what’s actually observable.
So what’s the way forward? A hybrid approach:
Human-style governance, machine-style control.
- Use human delegation concepts to assign human responsibility: define who is accountable and who can approve/stop.
- Use security engineering to control agents: scoped permissions, isolation, provenance, rate limits, kill switches, and lean on “verification” only in places where it can genuinely prove something.
If we get this right, delegation becomes scalable. If we get it wrong, we create potential industrial-scale mishaps.
https://lnkd.in/ez5a3STx
#AI #AIAgents #SafetyEngineering #Governance #AgenticSystems
Contact
bruno.gentil@sherpaconsultingasia.com
© 2026. All rights reserved.
