Policy Brief  

How AI Fails Us

Researchers and funders should redirect focus from centralized autonomous general intelligence to a plurality of established and emerging approaches that extend cooperative and augmentative traditions as seen in successes such as Taiwan’s digital democracy project to collective intelligence platforms like Wikipedia.

A human and robot hand touch

The dominant vision of artificial intelligence imagines a future of large-scale autonomous systems outperforming humans in an increasing range of fields. This “actually existing AI” vision misconstrues intelligence as autonomous rather than social and relational. It is both unproductive and dangerous, optimizing for artificial metrics of human replication rather than for systemic augmentation, and tending to concentrate power, resources, and decision-making in an engineering elite.  Alternative visions based on participating in and augmenting human creativity and cooperation have a long history and underlie many celebrated digital technologies such as personal computers and the internet.  Researchers and funders should redirect focus from centralized autonomous general intelligence to a plurality of established and emerging approaches that extend cooperative and augmentative traditions as seen in successes such as Taiwan’s digital democracy project to collective intelligence platforms like Wikipedia. We conclude with a concrete set of recommendations and a survey of alternative traditions.

More from this Program

India & the Olympics of AI

Commentary

India & the Olympics of AI

Allen Lab Fellow Jeremy McKey reflects on India’s AI Impact Summit, exploring the theme of diffusion and the implications for sovereignty and democracy.

Transparency is Insufficient: Lessons From Civic Technology for Anticorruption

Commentary

Transparency is Insufficient: Lessons From Civic Technology for Anticorruption

Allen Lab Researcher David Riveros Garcia draws on his experience building civic technology to fight corruption in Paraguay to make the case that effective civic technology must include power and collective action in its design.

More on this Issue

India & the Olympics of AI

Commentary

India & the Olympics of AI

Allen Lab Fellow Jeremy McKey reflects on India’s AI Impact Summit, exploring the theme of diffusion and the implications for sovereignty and democracy.

The Ecosystem of Deliberative Technologies for Public Input

Additional Resource

The Ecosystem of Deliberative Technologies for Public Input

Ensuring public opinion and policy preferences are reflected in policy outcomes is essential to a functional democracy. A growing ecosystem of deliberative technologies aims to improve the input-to-action loop between people and their governments.

Ethical-Moral Intelligence of AI

Occasional Paper

Ethical-Moral Intelligence of AI

In a new working paper, Crocodile Tears: Can the Ethical-Moral Intelligence of AI Models Be Trusted?, Allen Lab authors Sarah Hubbard, David Kidd, and Andrei Stupu introduce an ethical-moral intelligence framework for evaluating AI models across dimensions of moral expertise, sensitivity, coherence, and transparency.