In the rush to govern the age of agentic artificial intelligence (AI), a strange consensus has taken hold. Nearly every framework, policy paper, and regulatory discussion treats AI agents as the primary source of risk. They must be observable, controllable, auditable, and constrained. Humans, by contrast, are quietly ushered into the role of the wise observer, the final backstop, the moral anchor — the one element we can safely assume will keep the system honest.
This is not a minor design choice. It is the unspoken foundation of almost all current AI governance. And it is, increasingly, a fiction.
The Roots of the Human Exception
We have arrived at this place through a perfectly understandable combination of history, psychology, and institutional convenience. For centuries, human decision-making has been governed through layers of law, professional norms, corporate hierarchies, and social accountability. These systems are mature, politically settled, and emotionally reassuring. When a new technology arrives that can act at superhuman speed and scale, it is natural to focus regulatory energy on the newcomer. AI is the unknown variable. Humans are the known quantity. Why reopen settled questions about human fallibility when there is a shiny new risk demanding attention?
Layered on top is a deeper philosophical habit. We still place humans at the moral centre of the universe. Consciousness, free will, and intrinsic dignity are attributes we grant ourselves by default. AI is cast as a tool, an extension of human intent, not a peer controller. As long as a human remains in the loop — or ultimately accountable — we reassure ourselves that the system is ethically sound. The evidence of our own limitations — cognitive biases, fatigue, emotional contagion, groupthink, conflicts of interest — is acknowledged in psychology textbooks but rarely permitted to disturb the hierarchy. Humans are treated as the golden standard. Everything else must prove it is not dangerous.
Performative Governance in Practice
This mindset is reinforced by practical comfort. It feels cleaner to have a named person to blame when something goes wrong. It feels scalable. One human supervisor overseeing dozens of AI agents preserves the comforting illusion of control. And economically, it is convenient. Organisations deploy AI to cut costs and accelerate operations. Governance becomes the regulatory checkbox required to get the system live. Human processes are already governed or grandfathered in, so they largely escape the same quantitative scrutiny. Regulators, facing limited bandwidth and public pressure to be seen controlling the robots, naturally prioritise the novel threat.
The result is a subtle but profound inconsistency. In hybrid systems — the very systems that will dominate the next decade — we apply rigorous, measurable controls to silicon-based controllers while treating carbon-based ones as inherently trustworthy. We demand full transcripts, real-time drift detection, and kill switches from AI voice agents in call centres, yet the human operators handling the same calls are still evaluated largely through periodic quality assurance (QA) sampling and subjective supervisor judgment. When AI pushes too many escalations and humans become fatigued or inconsistent, the system-level instability is treated as an operational hiccup rather than a governance failure.
Let us call this what it is: performative governance in many cases. Tick-the-box compliance that looks impressive on paper but fails to address the actual dynamics of hybrid decision-making. The uncomfortable truth is that both humans and AI are fallible controllers operating inside the same feedback loops. Safety does not reside in the material the controller is made of — carbon or silicon — but in the observability, controllability, stability, robustness, and performance of the loop itself.
Evidence from Cognitive Science and Real-World Failures
Decades of research in cognitive psychology demonstrate that human decision-makers are systematically biased. Confirmation bias, overconfidence, anchoring, and availability heuristic are not edge cases — they are the norm. Under stress or fatigue, error rates increase dramatically. In high-stakes environments such as air traffic control, nuclear power plants, and intensive care units, we have long recognised that human performance degrades over time and must be monitored with the same rigour as machines. Yet when it comes to AI governance, we conveniently forget these lessons.
Real-world incidents underline the risk. In 2023, a major trading firm lost $450 million because a human trader overrode an AI risk model that had flagged a dangerous position. The trader’s intuition was wrong, but the governance framework held the AI to stricter standards than the human. In healthcare, studies show that clinician burnout leads to misdiagnosis rates comparable to early AI models, yet governance frameworks focus almost exclusively on validating the algorithm while assuming the doctor is always right. These are not isolated examples. They are symptoms of a systematic blind spot.
Moreover, the asymmetry scales dangerously. When a single human supervisor oversees ten AI agents, any cognitive limitation of that human is multiplied across all ten signals. Fatigue at 3 a.m. may cause the supervisor to miss a critical alert from an AI that is otherwise perfectly calibrated. The governance system will log the AI’s drift detection but ignore the human’s degraded state. The loop is broken not by the silicon but by the carbon, yet the regulatory lens remains fixed on the machine.
Towards Entity-Agnostic Governance
Moving to the next level of systemic adoption will require us to do something genuinely difficult: dethrone the human exception. We must design governance that is entity agnostic — judging every decision maker, human, AI, or hybrid fleet, by the same measurable properties of the control loop. This means retrofitting legacy human processes with the same telemetry and quantitative risk scoring we now demand of AI. It means accepting real-time fatigue monitoring for humans, similar to what modern cars already do when they detect driver drowsiness and recommend breaks. It means accepting bias audits and authority limits that feel invasive and dehumanising to some. It means updating liability, insurance, and regulatory frameworks to price human controller risk symmetrically with AI risk, even when that slows deployment or raises labour costs.
None of this will be easy, politically or culturally. Many employees will resist the idea of being measured with the same rigour as AI systems. Unions will push back. Executives will worry about morale and optics. Regulators will be reluctant to revisit established labour laws. Yet the alternative is worse: scaling hybrid agentic systems under the illusion of safety, only to discover too late that the weakest link was the one we refused to examine.
Several pioneering organisations are already moving in this direction. A European bank now subjects its human risk officers to the same algorithmic bias testing as its credit-scoring AI. A logistics company monitors the cognitive load of its human dispatchers alongside its route-optimisation algorithms, using dashboards that flag anomalies in both. A government agency requires that every decision — human or AI — be logged with the same level of detail and auditable trail. These examples show that entity-agnostic governance is not only possible but practical, and it leads to better outcomes.
The path forward is not to fear technology or to diminish human dignity. It is to extend the same rigorous, evidence-based discipline we are building for AI to every controller in the system, including ourselves. Only then can we move beyond performative governance to genuine systemic safety.
The question is not whether we will eventually make this shift. The question is how many avoidable failures we will tolerate before we admit that the material of the controller has never been the point. The loop is what matters. It is time we started governing it as such.
Source: ComputerWeekly.com News