Research

Paper

AI LLM February 25, 2026

The economic alignment problem of artificial intelligence

Authors

Daniel W. O'Neill, Stefano Vrizzi, Noemi Luna Carmeno, Felix Creutzig, Jefim Vogel

Abstract

Artificial intelligence (AI) is advancing exponentially and is likely to have profound impacts on human wellbeing, social equity, and environmental sustainability. Here we argue that the "alignment problem" in AI research is also an economic alignment problem, as developing advanced AI inside a growth-based system is likely to increase social, environmental, and existential risks. We show that post-growth research offers concepts and policies that could substantially reduce AI risks, such as by replacing optimisation with satisficing, using the Doughnut of social and planetary boundaries to guide development, and curbing systemic rebound with resource caps. We propose governance and business reforms that treat AI as a commons and prioritise tool-like autonomy-enhancing systems over agentic AI. Finally, we argue that the development of artificial general intelligence (AGI) may require a new economics, for which post-growth scholarship provides a strong foundation.

Metadata

arXiv ID: 2602.21843
Provider: ARXIV
Primary Category: econ.GN
Published: 2026-02-25
Fetched: 2026-02-26 05:00

Related papers

Raw Data (Debug)
{
  "raw_xml": "<entry>\n    <id>http://arxiv.org/abs/2602.21843v1</id>\n    <title>The economic alignment problem of artificial intelligence</title>\n    <updated>2026-02-25T12:22:46Z</updated>\n    <link href='https://arxiv.org/abs/2602.21843v1' rel='alternate' type='text/html'/>\n    <link href='https://arxiv.org/pdf/2602.21843v1' rel='related' title='pdf' type='application/pdf'/>\n    <summary>Artificial intelligence (AI) is advancing exponentially and is likely to have profound impacts on human wellbeing, social equity, and environmental sustainability. Here we argue that the \"alignment problem\" in AI research is also an economic alignment problem, as developing advanced AI inside a growth-based system is likely to increase social, environmental, and existential risks. We show that post-growth research offers concepts and policies that could substantially reduce AI risks, such as by replacing optimisation with satisficing, using the Doughnut of social and planetary boundaries to guide development, and curbing systemic rebound with resource caps. We propose governance and business reforms that treat AI as a commons and prioritise tool-like autonomy-enhancing systems over agentic AI. Finally, we argue that the development of artificial general intelligence (AGI) may require a new economics, for which post-growth scholarship provides a strong foundation.</summary>\n    <category scheme='http://arxiv.org/schemas/atom' term='econ.GN'/>\n    <category scheme='http://arxiv.org/schemas/atom' term='cs.CY'/>\n    <published>2026-02-25T12:22:46Z</published>\n    <arxiv:primary_category term='econ.GN'/>\n    <author>\n      <name>Daniel W. O'Neill</name>\n    </author>\n    <author>\n      <name>Stefano Vrizzi</name>\n    </author>\n    <author>\n      <name>Noemi Luna Carmeno</name>\n    </author>\n    <author>\n      <name>Felix Creutzig</name>\n    </author>\n    <author>\n      <name>Jefim Vogel</name>\n    </author>\n  </entry>"
}