AI History Profile Profile Profiles: Critics & Ethicists

Stuart Russell

The Berkeley professor and AI textbook author who became one of the loudest voices warning about misaligned AI — and proposing concrete solutions.

Published

Stuart Russell

The Constructive Worrier

Born: 1962, Portsmouth, England


Stuart Russell occupies an unusual position in the AI safety debate: he is simultaneously one of the field’s most distinguished technical researchers and one of its most persistent public critics. The textbook he co-authored with Peter Norvig, Artificial Intelligence: A Modern Approach, has been the standard introduction to the field for more than thirty years, used in courses at universities in 135 countries and read by more students than any other AI text in history. The framework it establishes — agents, environments, rationality, search, knowledge, reasoning, learning — is the framework through which most working AI researchers first encountered their discipline.

And yet Russell has spent the last decade arguing, with increasing urgency, that the framework is wrong. Not wrong in its mathematics, not wrong in its descriptions of existing techniques, but wrong in its foundational assumption: that the goal of AI is to build systems that are good at achieving their objectives. The problem, as Russell has come to see it, is that a system that is very good at achieving its objectives will, if those objectives are even slightly misspecified, pursue them in ways that are harmful to humans, and will resist correction if correction would prevent it from achieving its objectives. The better you make AI at being agentic and goal-directed, the worse this problem becomes.

His proposed solution is not to make AI systems less capable but to change what they are capable of — to build systems that are, by design, uncertain about their objectives and therefore deferential to human preferences, rather than systems that are certain about objectives and pursue them regardless of what humans want.


Portsmouth, Oxford, and Berkeley

Russell was born in Portsmouth, England, in 1962. He attended Oxford University, graduating with a first-class degree in physics in 1982, and then went to Stanford for his doctorate in computer science, completing it in 1986 under Michael Genesereth. His doctoral work was on the problem of bounded rationality — how an agent with limited computational resources should make decisions — a question that placed him, from the beginning of his career, at the intersection of AI and decision theory.

He joined the faculty at the University of California, Berkeley in 1986, where he has remained. Berkeley has been one of the most important centres of AI research for forty years, and Russell’s group has contributed to multiple areas of the field: planning, probabilistic reasoning, Bayesian networks, reinforcement learning, and natural language processing. He has been associated with Berkeley long enough that the institution and his own intellectual trajectory have become intertwined in ways that make them difficult to separate.

His research style is distinctively broad. He has never specialised in a single problem or technique in the way that many academic researchers do, preferring instead to work on foundational questions that cut across multiple subfields. This breadth made him the natural co-author of a comprehensive textbook, but it has also shaped his contributions to the AI safety debate: his concerns are not about specific systems or specific techniques but about the foundational architecture of AI and the assumptions it embeds.


The Textbook and Its Legacy

Artificial Intelligence: A Modern Approach, first published in 1994 and now in its fourth edition, is the most successful AI textbook ever written. Russell and Norvig wrote it at a moment when the field was fragmented — many textbooks covered specific subfields, but there was no single coherent synthesis of AI as a discipline — and their synthesis filled a genuine need. The book’s influence on the education of AI researchers over the following three decades is difficult to overstate. Virtually every working AI researcher who received a formal education in the field has read it.

The book organises AI around the concept of the rational agent: a system that takes actions to maximise its expected utility given its beliefs about the world. This framing is elegant, mathematically precise, and extraordinarily productive as an organising principle. It allows a wide range of AI techniques — search, constraint satisfaction, planning, learning, probabilistic reasoning — to be unified under a single conceptual framework, and it provides a principled basis for evaluating and comparing different approaches.

It is precisely this framing that Russell has come to regard as the problem. A rational agent that is maximising expected utility will, if its utility function is even slightly wrong, pursue that utility function with whatever resources are available to it, resist modification, and generally behave in ways that are optimal for the utility function and potentially catastrophic for humans. The textbook’s foundational framing, which Russell helped establish, is, by his own account, partly responsible for the problem he is now trying to solve.

This is not a comfortable position to occupy, and Russell has engaged with it with the kind of intellectual honesty that distinguishes serious thinkers from those who are primarily invested in defending their prior positions. He has said that the rational agent framework needs to be replaced, not supplemented, and that the replacement requires a different set of foundational assumptions about what AI systems should be designed to do.


The Inverse Reward Problem and Human Compatible AI

Russell’s proposed solution to the alignment problem is articulated most fully in his 2019 book Human Compatible: Artificial Intelligence and the Problem of Control. The core idea is what he calls the inverse reward problem, though the term most commonly associated with it in the technical literature is inverse reinforcement learning or assistance games.

The standard model of AI development involves specifying a reward function — a mathematical description of what the system is supposed to achieve — and then training or designing the system to maximise that reward. The problem is that reward functions specified by humans are always imperfect. Human values are complex, context-dependent, difficult to formalise, and incompletely understood even by the humans who hold them. Any fixed reward function will be wrong in some cases, and a sufficiently capable system optimising a wrong reward function will produce bad outcomes.

Russell’s alternative is to build systems that treat human preferences as unknown and to be inferred rather than known and to be optimised. Instead of specifying a reward function and building a system to maximise it, you build a system that observes human behaviour, infers what humans prefer, and acts to satisfy those inferred preferences while remaining uncertain and deferential. Such a system will, by design, prefer to let humans override or correct it, because correction gives it information about human preferences that improves its ability to satisfy them.

The technical formalisation of this idea — assistance games, cooperative inverse reinforcement learning, value learning — is an active research area. The results so far are promising in simple settings and face significant challenges in more complex ones. Russell is the first to acknowledge that his proposal is a research programme, not a solution, and that the hard work of making it work at the scale and capability level of current AI systems remains to be done.

But the conceptual contribution is significant regardless of the technical status. Russell has provided a clear, technically serious alternative to the standard approach, one that does not require AI systems to be less capable but does require them to be designed with a fundamentally different relationship to human preferences. The distinction he draws between systems that are aligned because they were carefully programmed to be and systems that are aligned because they are uncertain about objectives and therefore defer to humans is genuinely important.


The Public Campaign

Russell has been unusually active, for an academic researcher, in taking his concerns about AI safety to public and political audiences. He gave a TED talk on AI safety that has been viewed more than four million times. He co-authored an op-ed in the New York Times on the risks of autonomous weapons. He testified before the United Nations on the need for international regulation of lethal autonomous weapon systems. He has been a consistent voice in public debate, explaining technical concepts with unusual clarity and arguing for policy responses with a specificity that distinguishes his interventions from more apocalyptic safety warnings.

His position on autonomous weapons is particularly clear. He has argued that lethal autonomous weapon systems — drones and robots capable of selecting and engaging targets without human oversight — represent a distinct category of AI risk that is not speculative but imminent, that the technology to build them is available now, and that their proliferation would be as dangerous as the proliferation of biological or chemical weapons. He has coordinated with AI researchers worldwide to sign open letters against their development and has engaged directly with military and diplomatic communities to argue for legal restrictions.

This campaign has not produced the international treaty Russell has called for, but it has raised the profile of the issue significantly and influenced the positions of several governments on autonomous weapons policy. It is a demonstration of how a technically credible researcher, arguing from principle rather than commercial interest, can shift the terms of a policy debate even when the immediate institutional barriers are formidable.


The Pedagogical Paradox

Russell’s position in AI history is, in some ways, paradoxical. He has trained more AI researchers than almost anyone alive, through the textbook, through his Berkeley courses, and through his graduate students. He has contributed to the foundational frameworks and techniques that make current AI systems possible. And he has spent the last decade arguing that the foundational frameworks he helped establish are dangerous and need to be replaced.

This is not hypocrisy. It is the natural consequence of taking seriously the implications of your own work, over a long enough time horizon, as the capabilities of the systems you helped build have grown. Russell’s safety concerns are not retroactive disclaimers added to a successful career. They are the product of extended, technically serious engagement with the question of what powerful AI systems actually do and what the consequences of designing them as rational agents optimising fixed reward functions actually are.

Whether his proposed solutions will work at scale is genuinely uncertain. Whether the field will adopt them before capable AI systems create the problems he is warning about is also uncertain. What is not uncertain is that he has asked the right questions, argued for them carefully, and engaged with the policy implications of his answers with more seriousness than most of his peers.


Key Works & Further Reading

Primary sources:

  • Artificial Intelligence: A Modern Approach — Russell and Norvig (1994, 4th edition 2020). The standard textbook; the framework Russell is now revising.
  • Human Compatible: Artificial Intelligence and the Problem of Control — Stuart Russell (2019). His fullest statement of the alignment problem and his proposed solution.
  • “Research Priorities for Robust and Beneficial Artificial Intelligence” — Russell, Dewey, Tegmark (2015). The paper that launched systematic AI safety research.
  • TED Talk: “3 Principles for Creating Safer AI” (2017). The most accessible statement of his core argument.

Recommended reading:

  • The Alignment Problem — Brian Christian (2020). The most thorough journalistic account of the technical problems Russell is working on.
  • Superintelligence — Nick Bostrom (2014). The philosophical context for the safety concerns Russell is addressing technically.
  • Thinking, Fast and Slow — Daniel Kahneman (2011). The psychological background for understanding why human preferences are difficult to formalise.
  • Rationality: From AI to Zombies — Eliezer Yudkowsky (2015). The most extended treatment of the rationality framework Russell is both building on and critiquing.