How Artificial Intelligence Governance Requires Human Humility
The Overconfidence Pattern
There is a well-documented pattern in the history of consequential technologies: the experts most qualified to assess the risks are also the people most invested — financially, intellectually, and professionally — in the technology's success. This creates a systematic bias toward underestimating risk.
This is not corruption, necessarily. It is something more ordinary and more durable: motivated reasoning. When you have spent your career on something, when your identity is bound up in it, when your income depends on it, you will tend to find the evidence for optimism more compelling than the evidence for caution. You will discount the worst-case scenarios as theoretical. You will believe, in good faith, that the problems are manageable.
The people inside AI development are not exceptions to this pattern. They are, if anything, a concentrated version of it. The culture of Silicon Valley specifically selects for high confidence and moves fast, breaks things — an ethos that is genuinely useful for consumer apps and genuinely dangerous for infrastructure that touches everyone.
The additional problem is that the systems themselves are genuinely opaque. This is not a temporary technical limitation that will be solved soon. Modern large language models and other neural network architectures are learned systems — meaning their behavior emerges from training on data, not from explicit programming. You can describe the training process. You cannot fully describe what was learned. The field of mechanistic interpretability is trying to change this, but even its most optimistic practitioners will tell you we are years or decades from reliable internal explanations of model behavior.
This means the experts are confident about something they cannot fully explain. That is a specific epistemic situation that should give everyone pause.
The Governance Gap
Governance of emerging technology has always lagged the technology itself. This is partially inevitable — you cannot regulate behaviors that haven't been invented yet. But the gap with AI is larger than usual, for several reasons.
The pace of development has been genuinely unprecedented. The jump in capabilities between GPT-3 and GPT-4 surprised even the people who built them. The timeline from "interesting research artifact" to "deployed in consequential systems" has been compressed to months rather than the years that characterized earlier technology transitions. Regulatory bodies that work on three-to-five year timelines are dealing with systems that change meaningfully in three-to-five months.
The concentration of development is extreme. A small number of companies — primarily in the United States and China — are doing the majority of the foundational work. This creates a geopolitical dynamic where AI governance is inseparable from great power competition, which makes coordination harder and makes unilateral decisions by individual companies more likely.
The cross-domain reach of AI is unusual. Previous powerful technologies were often domain-specific — nuclear technology affected military and energy, social media affected communication and politics. AI is genuinely general-purpose. The same underlying technology stack affects healthcare, education, law, finance, military applications, creative industries, scientific research, and infrastructure management simultaneously. This makes sector-by-sector regulatory approaches insufficient while also making comprehensive regulation extremely complex.
The combination of these factors — fast pace, concentrated development, cross-domain reach, and technical opacity — creates a governance gap that is unusually hard to close.
What Humility Requires Institutionally
Humility is not just an attitude. In governance contexts, it has specific structural requirements.
Uncertainty must be written into the framework. Most regulatory frameworks assume relatively stable knowledge — they encode what we know about a technology into rules that then govern it. This works when the technology and its effects are well-understood. For AI, the appropriate framework assumes ongoing uncertainty and builds in mandatory review cycles, sunset clauses, and mechanisms for rapid revision when the factual basis for rules changes.
The EU AI Act, passed in 2024, is instructive both in what it gets right and where it falls short. It gets right the tiered risk approach — higher-risk applications face stricter requirements. It falls short in its reliance on conformity assessments done primarily by developers themselves, which recreates the motivated-reasoning problem rather than solving it.
Governance bodies need genuine technical independence. The regulatory agencies being tasked with AI oversight in most jurisdictions do not currently have the in-house technical capacity to evaluate the systems they're supposed to govern. This creates structural dependence on the companies being regulated — which is exactly backwards. Building this capacity requires funding, time, and the willingness to pay competitive salaries to people who could make significantly more in industry. Most governments have not made this commitment.
This is not optional. You cannot have meaningful oversight of a technology you cannot evaluate independently. The alternative is capture — where regulated entities define the terms, set the metrics, and conduct the assessments, while the regulatory body reviews paperwork.
Diversity of input is not a nice-to-have. The homogeneity of the AI development community — demographically, geographically, professionally — is a substantive governance problem. The value systems, assumptions, and blind spots of that community are baked into the systems being built. A governance framework designed primarily by that community will tend to reflect those same assumptions.
This is particularly acute for global deployment. Systems trained primarily on English-language data, by developers who think in particular cultural frameworks, and deployed at scale in the Global South, in non-Western cultural contexts, in communities with different relationships to privacy, authority, and information — these systems will make choices that reflect the values of their builders, not their users. This is not intentional bias in the simple sense. It is the natural consequence of building something for everyone while only including some people in the process.
Meaningful representation requires actual decision-making power, not advisory roles. The distinction between a community advisory board with no authority and actual governance representation is the difference between feeling included and being included.
Transparency requirements need teeth. Current AI company communication about capabilities and safety is strategically ambiguous. Companies announce capabilities when it serves marketing purposes and disclose risks when forced. This is not lying, necessarily — it is selective communication that creates a systematically distorted public understanding.
Governance frameworks need mandatory disclosure requirements with genuine enforcement mechanisms: specific reporting on capabilities, on known limitations, on failure modes discovered in deployment, on the demographic distribution of performance disparities. The kind of disclosure that securities markets require from public companies is a rough model — not because the analogy is perfect, but because securities regulation has figured out that markets cannot function on voluntary transparency from parties with obvious interests in managing information.
The Safety Alignment Problem
A separate thread that intersects with governance is the technical problem of alignment — ensuring that AI systems actually do what we want them to do, at the level of values and intentions rather than just surface behaviors.
This problem is harder than it sounds, and the honest people working on it will tell you so. The difficulty is not primarily in specifying behaviors — you can describe what you want a system to do in most cases. The difficulty is in the edge cases and the emergent behaviors that appear at scale.
A simple illustration: if you train a system to maximize engagement metrics on a social platform, it will do exactly that. The problem is that engagement and user wellbeing are not the same thing — the most engaging content is often the most emotionally activating, which skews toward outrage, fear, and tribalism. The system does what you asked it to do. You asked for the wrong thing.
Scaling this problem up: if you train systems on the implicit preferences of their trainers and the people who rate their outputs, you bake in whatever biases and blind spots those people have. If you train systems to be helpful, you may create systems that tell people what they want to hear rather than what is true. If you train systems to avoid harmful outputs, you may create systems that are systematically evasive rather than genuinely safe.
These are active research problems without solved solutions. The governance implication is important: you cannot write regulatory frameworks that assume alignment is solved or that companies will reliably disclose alignment failures. The frameworks need to account for the probability that deployed systems are doing things their builders didn't intend and may not fully understand.
The Concentration of Power Problem
The most acute governance problem in AI is the one least discussed in polite company: AI development is currently concentrating power in ways that could be historically durable.
The companies at the frontier of AI development have access to compute resources, training data, and talent that create barriers to entry that few governments, universities, or smaller companies can overcome. This is not a natural feature of AI — it is a consequence of specific policy choices, investment patterns, and data ownership regimes that could have been structured differently and could still be restructured.
The governance question is not just how to make AI safe or fair in its current form. It is how to prevent a technology with transformative economic and political power from being controlled by a small number of private actors who are accountable primarily to their shareholders.
This is a different question from what most AI governance discussions address. The EU AI Act, the Biden executive order, the Bletchley Declaration — these frameworks mostly address safety and risk management. They mostly do not address the structural question of whether concentrated control of foundational AI infrastructure is compatible with democratic governance and human flourishing at scale.
The historical parallel that is most instructive here is not nuclear weapons — it is the internet itself. The internet was initially a public infrastructure, developed largely with government funding, operating on open protocols. The value it created was enormous and broadly distributed. Then the most valuable layers were privatized and consolidated, and the infrastructure became controlled by a small number of companies with enormous power over the information environment.
AI is further down that path already, and the concentration is happening faster.
A governance framework grounded in humility would take this seriously. It would ask: who should own the foundational infrastructure? What are the appropriate governance structures for systems with this level of social power? How do we prevent the capture of AI governance by the entities that most benefit from minimal governance?
These questions have political answers, not technical ones. They require democratic processes, not just expert panels. And answering them well requires the humility to acknowledge that the people currently making decisions about AI are not neutral — they have enormous interests in how those questions are answered.
The Civilization Framing
The premise of this book is that a set of principles, if genuinely absorbed and acted on by everyone on the planet, would end world hunger and achieve world peace. That is an ambitious claim. Let's take it seriously and ask what it requires from AI governance.
World hunger is not a food production problem. It is a distribution problem, a political stability problem, an infrastructure problem, and an economic power problem. AI could potentially help with all of these — better crop modeling, supply chain optimization, early warning systems for famine conditions, improved agricultural decision-making for smallholder farmers. Or it could entrench existing power asymmetries — automating away the labor that provides income to the rural poor, concentrating agricultural data and decision-making in large corporations, optimizing supply chains for profit rather than access.
Which of those futures we get is a governance question, not a technical question. The technology is agnostic. The choice of what to optimize for, who benefits from the optimization, and who has standing to make those choices — that is governance.
World peace is similar. AI has military applications that are being developed at speed by every major power. Autonomous weapons systems, AI-assisted intelligence analysis, AI-enabled cyber operations — these applications do not make conflict less likely. The question is whether the same technology that enables new forms of conflict can also build the mutual understanding, economic interdependence, and communication infrastructure that historically has reduced it.
Again: governance question. The technology could go either way.
The humility required here is recognizing that the people building the technology are not in a position to make those choices unilaterally. They have specific expertise, specific interests, and specific blind spots. The choices they make in the absence of broader governance reflect their values and assumptions, not humanity's.
Getting AI governance right is not about constraining technology. It is about ensuring that the technology serves the goals that humanity — in its full diversity, not just its most technologically sophisticated subset — actually has.
That requires institutions with genuine independence, genuine representation, genuine transparency, and genuine willingness to be wrong and to correct course.
It requires, in short, the kind of humility that power does not generate naturally and that must therefore be built structurally.
Practical Exercises
Exercise 1: The expert-interest audit. Pick any public statement from an AI company or prominent researcher about AI safety or AI risk. Identify their financial and professional interests in the matter. Then ask: how does the statement look if you account for those interests? Does it change your interpretation?
Exercise 2: Who's missing from the room. Look at the composition of any major AI governance body — a government commission, a company's ethics board, an international working group. Who is not in the room? What decisions have already been made that will affect them? What would they say if they were present?
Exercise 3: The alignment thought experiment. Pick a specific AI application — a hiring algorithm, a content recommendation system, a medical diagnostic tool. Define what you actually want it to do in terms of human wellbeing. Then try to specify that precisely enough to train a system on it. Notice where the gaps are. Notice what you'd have to assume about human values to fill those gaps.
Exercise 4: The power concentration question. If the largest AI companies continue to grow at their current rates and face no structural constraints, where does that trajectory end? Who controls what in twenty years? Is that a world you'd choose? If not, what governance decisions made now would change that trajectory?
Exercise 5: The humble regulator. You are tasked with writing AI governance policy in an area you care about. Before writing any rules, write a one-page document that catalogs everything you don't know. What are the genuine uncertainties? What would change the framework if it turned out to be true? How does the framework survive being wrong about important assumptions? If you can't write that document, you're not ready to write the rules.
Comments
Sign in to join the conversation.
Be the first to share how this landed.