The 80/20 Utopia Gamble
Abstract
This paper examines the 80/20 Utopia Gamble in AI ethics, where optimisation normalises sacrificing minorities for aggregate gains. Through utilitarian, deontological, Rawlsian, capability, and political-economy analysis, it argues this logic is unstable, unjust, and dangerous, proposing justice-constrained objectives, participatory governance, and explicit rejection of catastrophic-risk trade-offs in AI systems.
Introduction: The quiet normalisation of sacrifice
Introduction: The quiet normalisation of sacrifice
The “80/20 Utopia Gamble” names an uncomfortable intuition that is beginning to crystallise in contemporary AI ethics: powerful actors may be willing to accept serious, even catastrophic, harms to a minority so long as technological systems deliver visible benefits to a majority. In recent interviews, Tristan Harris describes conversations with frontier AI leaders who privately accept significant probabilities of civilisation-level harm or extinction as the price of racing to artificial general intelligence (AGI), while publicly emphasising visions of abundance, health, and prosperity for “everyone.” This is framed explicitly as an “80/20” calculation: gamble an extraordinary upside for 80% of humanity (or more) against grave risks borne by some “20%,” including future generations or those structurally exposed to failure.
The distinctive ethical problem, however, is not only in the numbers. It is that this gamble is being made implicitly, embedded in optimisation objectives, incentive structures, and institutional competition, rather than through public, contestable moral deliberation. The danger is that unexamined optimisation, “runaway optimisation,” in Harris’s language, quietly becomes the de facto moral framework for AI. This paper argues that the 80/20 Utopia Gamble is an unstable and ethically indefensible basis for AI development when examined through utilitarian, deontological, and justice-based lenses, and that alternative frameworks, especially Rawlsian and capability-based approaches, better capture what is at stake.
The seduction and failure of naïve utilitarianism
The seduction and failure of naïve utilitarianism
At first glance, the 80/20 framing appears to be a straightforward application of utilitarian reasoning. Classical utilitarianism defines right action as that which maximises overall happiness or well-being across all affected parties, the “greatest good for the greatest number.” In policy contexts, this often collapses into a cost–benefit calculus: if the aggregate benefits to many outweigh the aggregate harms to a few, the trade-off appears justified.
Large-scale AI deployment almost invites this mindset. Models that increase economic productivity, reduce medical errors, or enhance safety in transportation can plausibly generate enormous aggregate benefits. Against this, harms such as increased surveillance, job displacement in specific sectors, or the systematic misclassification of some minority groups are often framed as regrettable but acceptable side-effects, “friction” that can be managed at the margins of an otherwise socially beneficial optimisation.
However, even on its own terms, this is a distorted utilitarianism.
- First, empirical work on algorithmic systems shows that harms to disadvantaged groups are routinely undercounted or invisible in the very metrics used to judge “overall benefit.” Health risk algorithms have been shown to allocate fewer resources to Black patients for the same level of need because they use historical spending as a proxy for illness severity, thereby encoding structural underinvestment as “lower risk.” Predictive policing systems intensify patrols in already over-policed neighbourhoods based on biased historical data, generating feedback loops that increase recorded crime and justify further surveillance. Tools used in criminal justice, employment, credit scoring, and welfare eligibility exhibit similar patterns of disparate impact and structural injustice.
- Second, contemporary utilitarian practice in technology often ignores the classic insight that impartiality is central: all suffering and flourishing must be weighed equally. In high-stakes AI, those designing and profiting from systems are generally not those most exposed to catastrophic downside risks, whether in terms of labour precarity, political repression, or existential vulnerability. As Harris notes, some AI leaders justify racing ahead by privately accepting non-trivial probabilities of human extinction, in part because they expect to be among the potential beneficiaries of radical life extension if things go well. This is not impartial utilitarianism; it is a self-serving, speculative expected-value calculation that discounts the interests of those who will never consent to being part of such an experiment.
- Third, when the gamble extends to existential risk, utilitarian arguments notoriously cut both ways. A tiny probability of infinite future value can dominate any finite present harm, but a tiny probability of irreversible extinction can equally swamp any finite benefit. There is no neutral “80/20” partition when the “20” includes the possibility of no future moral patients at all. Once this is acknowledged, the air of technocratic rationality around the gamble evaporates; what remains is a political choice about who is allowed to impose what risks on whom.
Deontological critique: persons as ends, not optimisation variables
Deontological critique: persons as ends, not optimisation variables
Deontological ethics takes a sharply different starting point. Kant’s account of morality grounds right action in duties owed to rational beings as ends in themselves, not in the promotion of aggregate welfare. The second formulation of the categorical imperative, “So act that you use humanity… always at the same time as an end, never merely as a means”, is particularly salient. To instrumentalise a subset of persons as the “20%” who may predictably lose agency, dignity, or even existence to secure utopian gains for others is a textbook violation of this principle.
Many contemporary AI practices already strain this duty of respect. Rumman Chowdhury has described “moral outsourcing,” where institutions hand over consequential value-laden decisions, such as who is targeted for policing or who receives life-changing opportunities, to opaque models, then treat the algorithm’s output as a morally exculpating authority. Commentary on AI in organisational contexts similarly warns that when optimisation functions and training data stand in for explicit moral reasoning, humans retain power while shedding perceived responsibility for outcomes. In such regimes, individuals are not engaged as autonomous agents capable of contesting the norms that govern them; they are treated as datapoints to be classified, nudged, or excluded.
This matters even if the systems appear beneficial “on average.” For example, recidivism risk scores and credit scoring algorithms that systematically mis-classify some groups may be defended because they improve public safety or economic efficiency overall. A deontological lens, however, insists that wrongful discrimination, manipulation, or denial of due process cannot be morally redeemed by aggregate benefits. Rights to equal respect, fair treatment, and participation in decisions that shape one’s life impose side-constraints on what may be done in pursuit of utopian outcomes.
From this perspective, the 80/20 Utopia Gamble is not merely a risky bet; it encodes a willingness to treat a predictable minority as morally fungible. That stance is irreconcilable with any ethics that takes inviolable rights and respect for persons as foundational.
Justice as fairness: the wrong 20%
Justice as fairness: the wrong 20%
Rawlsian theories of justice deepen this critique by shifting attention from isolated decisions to the basic structure of social institutions. Rawls’s “justice as fairness” holds that the principles governing social and economic inequalities are just only if they would be chosen behind a “veil of ignorance,” where no one knows their own social position, and if any inequalities work to the greatest benefit of the least advantaged.
Recent work in AI ethics explicitly operationalises Rawlsian ideas in algorithm design and assessment. Researchers have proposed fairness objectives that interpolate between utilitarian maximisation of aggregate utility and Rawlsian maximin, prioritising the worst-off group’s outcomes in model training and evaluation. Others argue that because AI increasingly shapes access to opportunities, social recognition, and political power, it must be governed by principles of distributive justice appropriate to the “basic structure” of a digital society.
Viewed through this lens, the 80/20 Utopia Gamble fails twice. First, behind a veil of ignorance, no rational agent would endorse a social order in which a fixed 20% of persons could foreseeably suffer severe losses of agency, dignity, or life so that an 80% could enjoy enhanced convenience or prosperity, especially when membership of the sacrificial group is tightly correlated with existing axes of structural disadvantage. Second, the difference principle demands that permissible inequalities demonstrably improve the position of the least advantaged; yet many AI deployments, from welfare fraud detection to predictive policing and biometric border control, predictably worsen the situation of already oppressed groups for the sake of efficiency or security gains enjoyed primarily by others.
Rawlsian analyses of AI thus support a simple but powerful reformulation: the relevant question is not “What fraction of people benefit?” but “What happens to the least advantaged under this system?” Any AI “utopia” that consigns a structurally identifiable 20% to worsened prospects, or gambles with their continued existence, is unjust by design.
Capabilities and relational ethics: when “utopia” fails basic thresholds
Capabilities and relational ethics: when “utopia” fails basic thresholds
Capability theories, associated with Amartya Sen and Martha Nussbaum, offer another way to see why aggregate improvements can still be ethically intolerable. Instead of focusing on utility or resources, these approaches ask whether each person has the real freedom to achieve a set of central human capabilities, such as bodily health, bodily integrity, practical reason, affiliation, and control over one’s environment. A social order is just if it secures at least a threshold level of these capabilities for everyone.
AI systems that enhance productivity or convenience for a majority while eroding basic capabilities for a minority, for example, by enabling ubiquitous surveillance, automating welfare sanctions, or entrenching discriminatory search and recommendation systems, cannot be justified by pointing to large net benefits. If 20% of people lose the capability for meaningful political participation, secure employment, privacy, or bodily integrity as a consequence of optimisation choices, the resulting “utopia” fails the minimal conditions of a decent society.
Relational and care-based ethics sharpen this point. Work on AI in healthcare and other intimate domains has highlighted how opaque, “black box” systems can undermine patient autonomy and relational trust, even when clinically effective. Care ethics emphasises responsibilities to maintain and repair relationships of dependence and vulnerability, not just to deliver outcomes. On this view, the 80/20 gamble corrodes the moral fabric of society by normalising the idea that some relationships, those with people in the sacrificial 20%, are expendable so long as macro-level metrics improve.
Political economy and “runaway optimisation”
Political economy and “runaway optimisation”
The 80/20 Utopia Gamble does not arise in a vacuum. It is a product of a particular political economy that Shoshana Zuboff terms “surveillance capitalism”: a new logic of accumulation built on extracting behavioural data, predicting and shaping human behaviour, and trading in “behavioural futures markets.” In this regime, digital architectures are designed to maximise engagement, scale, and behavioural control, often at the expense of privacy, autonomy, and democratic oversight.
Harris and the Center for Humane Technology argue that AI is following the same “dangerous playbook” that earlier social media platforms pioneered, with optimisation objectives aligned to engagement and market dominance rather than human flourishing. Algorithms are deployed at scale as what Cathy O’Neil calls “weapons of math destruction”: opaque, high-impact systems whose errors and biases disproportionately damage already vulnerable populations. Once embedded in hiring, policing, credit, and welfare systems, these models can quietly redistribute risk and opportunity along lines of race, class, and geography, even as the platforms that operate them report substantial aggregate gains in efficiency or profit.
Under these conditions, the 80/20 framing risks functioning as a legitimising narrative. It recasts structurally patterned harms as the unfortunate but rational by-product of an otherwise inevitable technological trajectory. The critical question, who selected the utility function; who captures the upside; who bears the downside; and who has standing to refuse the bargain, disappears behind technical language about optimisation and trade-offs.
From implicit gamble to explicit ethics
From implicit gamble to explicit ethics
The central normative insight of the 80/20 Utopia Gamble is diagnostic rather than prescriptive. It reveals that unexamined optimisation in AI tends to smuggle in a crude, partial utilitarianism that normalises sacrifice and obscures responsibility. Once surfaced, this logic can, and should, be rejected as a default moral framework.
Several directions follow for AI research, governance, and practice.
- Optimisation objectives should be constrained by justice-sensitive principles rather than left as purely technical or commercial design choices. Rawlsian and capability-based approaches provide concrete templates: algorithmic systems should be evaluated not only for average performance, but for their impact on the least advantaged and on each person’s ability to exercise key capabilities. This suggests adopting robust, maximin-like objectives in high-stakes domains, along with distributional metrics that make harms to minorities visible and non-negotiable.
- AI ethics work needs to move beyond high-level principles to institutional mechanisms that prevent moral outsourcing. Empirical reviews show convergence around transparency, fairness, non-maleficence, responsibility, and privacy in AI ethics guidelines, but implementation is often weak. Requirements for explainability, contestability, independent auditing, and meaningful human oversight in consequential systems can help ensure that no actor can plausibly say “the system decided; it was not my choice.”
- Those most at risk of being in the “20%” must have a substantive role in shaping AI systems and policies. Community participation in the design, deployment, and governance of AI, particularly in policing, welfare, healthcare, and migration control, is essential if distributive impacts are to be assessed and corrected in time. Otherwise, the gamble will predictably reproduce existing hierarchies under the guise of neutral optimisation.
- AI research communities should explicitly refuse framings that treat existential or catastrophic risks as acceptable collateral for speculative utopias. If leading developers privately assign non-trivial probabilities to civilisational collapse, proceeding as though the “expected value” is positive is not a responsible stance; it is a political decision to externalise unprecedented risks onto the rest of humanity and future generations. An ethics that centres duties, justice, and capabilities would instead require that such risks be reduced to a level compatible with secure, dignified lives for all before large-scale deployment is considered legitimate.
The most important question posed by the 80/20 Utopia Gamble is thus not whether AI can make 80% of people better off. It is whether AI will become the latest technology through which societies habituate themselves to the idea that some people, some communities, or some futures are expendable. A humane AI ethics must answer that question in the negative and redesign its systems, incentives, and institutions accordingly.
References
Rawls, J. (1999). A theory of justice (Rev. ed.). [Harvard University Press.](https://www.hup.harvard.edu/books/9780674000780) Kant, I. (2019). Groundwork of the metaphysics of morals (M. Gregor & J. Timmermann, Trans.). Cambridge University Press. [Cambridge University](https://www.cambridge.org/core/books/groundwork-of-the-metaphysics-of-morals/6F5A3A38F8C3E1AEBB7D62E9E5B2C5C6) Nussbaum, M. C. (2011). Creating capabilities: The human development approach. Harvard [University Press.](https://www.hup.harvard.edu/books/9780674072350) Barocas, S., & Selbst, A. D. (2016). Big data’s disparate impact. California Law Review, 104(3), 671–732. [California Law Review](https://www.californialawreview.org/print/big-datas-disparate-impact) Jobin, A., Ienca, M., & Vayena, E. (2019). The global landscape of AI ethics guidelines. Nature Machine Intelligence, 1(9), 389–399. [Nature Machine Intelligence](https://www.nature.com/articles/s42256-019-0088-2) Zhang, H., Lakkaraju, H., & Kleinberg, J. (2023). From utilitarian to Rawlsian designs for algorithmic fairness. Proceedings of the AAAI Conference on Artificial Intelligence, 37(11), 13929–13938. [AAAI Conference on Artificial Intelligence](https://ojs.aaai.org/index.php/AAAI/article/view/26525) Lepri, B., Oliver, N., Letouzé, E., Pentland, A., & Vinck, P. (2018). Fair, transparent, and accountable algorithmic decision-making processes. Philosophy & Technology, 31(4), 611–627. [Philosophy & Technology](https://link.springer.com/article/10.1007/s13347-017-0279-x) Zhang, B., Lemoine, B., & Mitchell, M. (2020). Measuring justice in machine learning. Proceedings of the ACM Conference on Fairness, Accountability, and Transparency, 263–273. [Fairness, Accountability, and Transparency](https://dl.acm.org/doi/10.1145/3351095.3372879) Zuboff, S. (2019). The age of surveillance capitalism: The fight for a human future at the new frontier of power. PublicAffairs. [Public Affairs Books](https://www.publicaffairsbooks.com/titles/shoshana-zuboff/the-age-of-surveillance-capitalism/9781610395694/) O’Neil, C. (2016). Weapons of math destruction: How big data increases inequality and threatens democracy. Crown. [The Crown](https://weaponsofmathdestructionbook.com/) Noble, S. U. (2018). Algorithms of oppression: How search engines reinforce racism. New York University Press. [New York University](https://nyupress.org/9781479837243/algorithms-of-oppression/) Selbst, A. D., & Barocas, S. (2018). The intuitive appeal of explainable machines. Fordham Law Review, 87(3), 1085–1139. [Fordham Law Review](https://ir.lawnet.fordham.edu/flr/vol87/iss3/4/) Schulz, D. (2021). Ethical machines: The human-centric use of artificial intelligence. Frontiers in Big Data, 4, 583151. [Frontiers in Big Data](https://www.frontiersin.org/articles/10.3389/fdata.2021.583151/full) Wong, K. (2024). Transparency and accountability in AI systems. Frontiers in Human Dynamics, 6, 1421273. [Frontiers in Human Dynamics](https://www.frontiersin.org/articles/10.3389/fhumd.2024.1421273/full) Harris, T. (2025). Technology ethicist Tristan Harris on the Diary of a CEO podcast [Transcript]. [Diary of a CEO](https://singjupost.com/technology-ethicist-tristan-harris-on-the-diary-of-a-ceo-podcast-transcript/) The Center for Humane Technology. (2025). The CHT perspective. [Center for Humane Technology](https://www.humanetech.com/the-cht-perspective)