Policy Brief  

Plural Publics

The authors highlight why we believe the problem of “plural publics” to be a core challenge of data governance, discuss existing tools that can help achieve it and a research agenda to further develop and integrate these tools.

By:

  • Divya Siddarth
  • Glen Weyl
  • Shrey Jain

Download the PDF

Plral Publics in white text on Teal background

Data governance is usually conceptualized in terms of “privacy” v. “publicity”. Yet a core feature of pluralistic societies is association, groups that share with each other, privately. These are a diversity of “publics”, each externally private but with the ability to coordinate and share internally. Empowering them requires tools that allow the establishment of shared communicative contexts and their defense against external sharing outside of context. The ease of spreading information online has challenged such “contextual integrity” and the rise of generative foundation models like GPT-4 may radically exacerbate this challenge. In the face of this challenge, we highlight why we believe the problem of “plural publics” to be a core challenge of data governance, discuss existing tools that can help achieve it and a research agenda to further develop and integrate these tools.

Related Resources

Crocodile tears: Can the ethical-moral intelligence of AI models be trusted?

Open Access Resource

Crocodile tears: Can the ethical-moral intelligence of AI models be trusted?

Allen Lab authors Sarah Hubbard, David Kidd, and Andrei Stupu introduce an ethical-moral intelligence framework for evaluating AI models across dimensions of moral expertise, sensitivity, coherence, and transparency in their recently published paper, Crocodile Tears: Can the Ethical-Moral Intelligence of AI Models Be Trusted? in Springer AI & Ethics.

AI & Democracy: Perspectives from an Emerging Field

Additional Resource

AI & Democracy: Perspectives from an Emerging Field

The Allen Lab is proud to have contributed to this timely landscape report from The David & Lucile Packard Foundation mapping the emerging field of AI and democracy.

More on this Issue

Crocodile tears: Can the ethical-moral intelligence of AI models be trusted?

Open Access Resource

Crocodile tears: Can the ethical-moral intelligence of AI models be trusted?

Allen Lab authors Sarah Hubbard, David Kidd, and Andrei Stupu introduce an ethical-moral intelligence framework for evaluating AI models across dimensions of moral expertise, sensitivity, coherence, and transparency in their recently published paper, Crocodile Tears: Can the Ethical-Moral Intelligence of AI Models Be Trusted? in Springer AI & Ethics.

AI & Democracy: Perspectives from an Emerging Field

Additional Resource

AI & Democracy: Perspectives from an Emerging Field

The Allen Lab is proud to have contributed to this timely landscape report from The David & Lucile Packard Foundation mapping the emerging field of AI and democracy.