As a part of the Allen Lab’s Political Economy of AI Essay Collection, Sarah Hubbard explores alternative ownership and governance structures for artificial intelligence that may better serve the public interest.
An examination of the development of artificial intelligence (AI) reveals a growing concentration of corporate power and wealth at various layers of the AI stack. Today, nearly every stage in AI model development — from compute infrastructure to training data — is controlled by a handful of large technology companies. Very little public oversight exists when it comes to the development and governance of these systems.
The risks associated with concentrating AI development in monopolistic powers are generally well understood and have been widely discussed in existing literature.1 There is a well-established need to explore alternative ownership and governance structures that would better serve the public interest as it pertains to AI development. In this essay, we will examine one such alternative structure: the cooperative. Cooperatives are worker-owned and operated firms with a rich global history of community-oriented business practices that distribute control and capital. Many well-known firms, such as REI, Ocean Spray, Dairy Farmers of America, and the Associated Press, operate as cooperative businesses. Today, it is estimated that cooperatives employ nearly 10% of the world’s population.2 While they aren’t without challenges, cooperative models are structured to provide more equitable ownership and governance than investor-owned corporations.3
Below, we will discuss how such a structure can benefit the development and governance of AI while easing the growing concerns of various stakeholders: consumers wanting input into the systems dominating their lives, companies looking to set themselves apart by gaining consumer and regulator trust, and regulators aiming to prevent market monopolization.
Opportunities for Cooperative AI
Public Interest Infrastructure
Today, only a few companies, such as NVIDIA and TSMC, produce the graphics processing units (GPUs) required to train AI models at scale. These GPUs are primarily sold to a small number of large technology companies that have the resources to leverage them in compute offerings. Amazon AWS, Microsoft Azure, and Google Cloud are estimated to account for two-thirds of the entire cloud market,4 which has led to various antitrust investigations.5 Companies trying to develop their own AI models may spend millions of dollars to access the compute resources they need for training on these platforms.6 As a result, some of the emerging leaders in AI, namely OpenAI and Anthropic, have had to strike agreements with the large cloud providers,7 often in exchange for equity. This concentration of computational infrastructure by a few large technology companies has priced out many smaller global developers, created ecosystem lock-in, and made healthy competition in the cloud computing space incredibly difficult.
In an attempt to create an alternative to large corporate cloud infrastructure, projects such as the Co-op Cloud and the Commons Cloud have emerged,8 alongside burgeoning research into cooperative cloud computing.9 While these projects do not currently support the type and scale of compute infrastructure needed for the large-scale training of AI systems, they demonstrate a clear desire and need for this type of public infrastructure. Recent efforts to create a “public AI” option further validate the need for an alternative to existing corporate cloud providers.10 Bills such as the CREATE AI Act propose creating a National Artificial Intelligence Research Resource to provide compute infrastructure for AI researchers and students.11
A recent paper, “An Antimonopoly Approach to Governing AI”,12 recommends cooperative alternatives:
“At the cloud layer, the federal government could support the creation of a cooperative research-focused cloud, owned and operated by nonprofits, government, and universities to ensure sufficient compute and storage power for research into innovative, safe uses of AI — and without a shareholder profit motive. The federal government could also support the creation of a cooperative cloud for private companies, in which firms could train and operate models, and share in the ownership of the cloud, without fear that one of the big platforms will take their ideas or raise prices for the utility services they provide.”
While some progress has been made toward creating compute infrastructure in the public interest, considerable challenges remain due to the capital-intensive requirements needed to provide alternatives to the corporate cloud providers. Additional experiments in federated compute infrastructure and cooperative clouds, as well as continued exploration into federal funding models for public AI infrastructure, represent promising steps toward more cooperative AI development.
Data Cooperatives
AI systems require vast amounts of data for model training, much of which is sourced from the public internet, including user-generated media, forum discussions, and social media posts. Most individuals are unaware of how their data is being used and have little power to opt in or out of its collection. Recently, a wave of copyright lawsuits coming from artists, authors, and publications claim that these AI models and businesses are “built on the exploitation of copyrighted works.”13
On the other hand, companies training AI models need high-quality data to produce a high-quality model. After Google acquired Reddit data to train its large language model, Bard, many of the AI’s responses were shown to be inaccurate or otherwise flawed due to the poor-quality Reddit forum responses they referenced.14 This type of low-quality data can impact AI’s accuracy and relevance, leading to poor user experiences and ultimately negative business outcomes.
Data cooperatives offer an alternative to current extractive practices by aiming to shift the power from large corporations to the individual. They enable individuals to pool their data while retaining control over its use and collectively benefiting from its access. A data cooperative empowers end users to control who can access their data, what data they can access, and for what purpose. Furthermore, it allows end users to negotiate ownership of their data, including any financial benefits that might be distributed back to them. Hardjono and Pentland describe the need for data cooperatives in their piece on “Personal Data Management”15:
“Today we are in a situation where individual assets … people’s personal data … [are] being exploited without sufficient value being returned to the individual. This is analogous to the situation in the late 1800’s and early 1900’s that led to the creation of collective institutions such as credit unions and labor unions, and so the time seems ripe for the creation of collective institutions to represent the data rights of individuals. We have argued that data cooperatives with fiduciary obligations to members provide a promising direction for the empowerment of individuals through collective use of their own personal data.”
A leading example of this concept is a data trust called Superset,16 which is structured to allow members to contribute, govern, and be compensated for their data. Superset then negotiates on behalf of its members in a collaboration with Delphia, a customer of the data trust.17 Other collective data examples include Cohere’s Aya,18 which collected natural language data from thousands of people around the world to train a multilingual AI model. While Aya’s contributors aren’t financial stakeholders in the project, both the data set and AI model are fully open-source for anyone to download and use.
Explorations into cooperative data agreements also appear in academic research. A study titled “Worker-Owned Cooperative Models for Training Artificial Intelligence” explored having data workers (who often label and clean data for low wages) earn ownership of the AI systems they help create.19 Similarly, a recent paper, “An Economic Solution to Copyright Challenges of Generative AI,”20 proposed “a simple framework that appropriately compensates copyright owners for using their copyrighted data in training generative AI systems based on the cooperative game theory …” Through recognizing the dual challenges of companies needing high-quality, often copyrighted data and AI disrupting creative industries, the authors present a cooperative solution to meet both parties’ needs.
Other examples where data cooperatives have proven successful in the technology sector include the Driver’s Seat Cooperative,21 supported by The Rockefeller Foundation,22 which enabled gig workers to boost their pay by crowdsourcing market information and optimizing their time. While not directly related to AI development, it showcases how data cooperatives can successfully scale and deliver benefits in other technology-dominated sectors.
Collective Governance
Lessons from the social media age demonstrate the tension that exists between the profit motives of large technology companies and their ability to self-govern to avoid diffuse harms to society. Given that AI will continue to have outsized impact on global stakeholders, it is critical to explore methods for greater collective community governance over the development and deployment of these systems. Many are rightfully skeptical of large technology companies’ claims to be committed to the public good. For example, in 2019, the Business Roundtable released a pledge of corporations to “commit to lead their companies for the benefit of all stakeholders — customers, employees, suppliers, communities and shareholders.”23 However, a follow-up investigation demonstrated that there was no material change in how companies treated these stakeholders.24
A variety of different corporate governance models are being tested in the AI space today. Anthropic is structured as a public benefit corporation and introduced a Long-Term Benefit Trust,25 which appoints independent trustees to oversee the Anthropic board. OpenAI is a nonprofit that runs a for-profit entity but has experienced highly publicized governance turmoil and debate in the past year.26 Meta launched the Oversight Board,27 an independent entity that makes decisions on content moderation cases, which Meta is required to follow. While these are steps in the right direction, there should be further exploration of operating models that empower end users as stakeholders with decision-making power over AI systems.
While there have been some shifts in corporate governance, there must be a greater movement toward empowering people to collectively and cooperatively govern the AI systems that are impacting their lives.
Sarah Hubbard
Associate Director for Technology & Democracy, Allen Lab for Democracy Renovation
There are increasing civil society efforts that advocate for collectively governed AI. The Collective Intelligence Project is leading alignment assemblies,28 which aim to incorporate public input into the development of AI systems. Other methods, such as citizens’ assemblies,29 are also being tested. For example, in Belgium, 60 citizens were randomly selected to discuss AI and inform the larger EU policy strategy.30 Other advocacy groups like the Algorithmic Justice League are working toward “equitable and accountable AI.”31 Another emerging alternative is the concept of “Exit to Community,”32 where instead of an IPO or acquisition, a company instead transitions to cooperative community control. Efforts toward this end include a bid for X (formerly Twitter)33 and a recent proposal to purchase and decentralize TikTok.34
While there have been some shifts in corporate governance, there must be a greater movement toward empowering people to collectively and cooperatively govern the AI systems that are impacting their lives. As the public’s trust in institutions and large technology companies continues to decline, shifting power to the public might garner greater trust and buy-in.
Recommendations
Cooperative methods for AI development and governance present an alternative approach to the reigning paradigm of today’s corporate technology power. To support this future, a few key recommendations follow:
Efforts to promote greater cooperative methods within the AI development life cycle are most likely to gain traction in outlined areas such as cloud, data, and governance, where there is ongoing momentum and antitrust discussions.
While there are increasing experiments in corporate governance structures and public input mechanisms in AI, it is critical to ensure that these structures give actual power, enforceable oversight, and an economic stake for community stakeholders.
The pace of AI development raises the urgency for greater research and experimentation with cooperative AI paradigms, as well as policy interventions to ensure a healthy competitive market and innovation centered around the public interest. Immediate action is needed to prevent continued consolidation of power and wealth of large technology companies.
Exploring more cooperative paradigms stands to benefit all stakeholders: consumers gain more influence and an economic stake in the technology systems shaping their lives, businesses build trust with consumers and regulators while staying connected to public needs, and regulators facilitate a competitive marketplace with the possibility of improving AI’s societal impact. The rising tide of AI should lift all boats.
Political Economy of AI Essay Collection
Earlier this year, the Allen Lab for Democracy Renovation hosted a convening on the Political Economy of AI. This collection of essays from leading scholars and experts raise critical questions surrounding power, governance, and democracy as they consider how technology can better serve the public interest.
Malik, Sheheryar, Fabrice Huet, and Denis Caromel. “Cooperative cloud computing in research and academic environment using Virtual Cloud.” 2012 International Conference on Emerging Technologies (2012): 1-7. https://doi.org/10.1109/ICET.2012.6375445.
Narechania, Tejas N., and Ganesh Sitaraman “An Antimonopoly Approach to Governing Artificial Intelligence.” Vanderbilt Law Research Paper, no. 24-28 (2024), available at SSRN. https://doi.org/10.2139/ssrn.4597080.
Sriraman, Anand, Jonathan Bragg, and Anand Kulkari. “Worker-Owned Cooperative Models for Training Artificial Intelligence.” | CSCW ’17 Companion: Companion of the 2017 ACM Conference on Computer Supported Cooperative Work and Social Computing (2017): 311-314. https://dl.acm.org/doi/10.1145/3022198.3026356.
Wang, Jiachen T., Zhun Deng, Hiroaki Chiba-Okabe, Boaz Barak, and Weijie J. Su. “An Economic Solution to Copyright Challenges of Generative AI.” arXiv (2024). https://doi.org/10.48550/arXiv.2404.13964.
Bebchuk, Lucian A., and Roberto Tallarita. “Will Corporations Deliver Value to All Stakeholders?” Vanderbilt Law Review 75 (2022): 1031-1091. https://doi.org/10.2139/ssrn.3899421.
The year 2024 was dubbed “the largest election year in global history” with half the world’s population voting in national elections. Earlier this year, we hosted an event on AI and the 2024 Elections where scholars spoke about the potential influence of artificial intelligence on the election cycle– from misinformation to threats on election infrastructure. This webinar offered a reflection and exploration of the impacts of technology on the 2024 election landscape.
Earlier this year, the Allen Lab for Democracy Renovation hosted a convening on the Political Economy of AI. This collection of essays from leading scholars and experts raise critical questions surrounding power, governance, and democracy as they consider how technology can better serve the public interest.
As a part of the Allen Lab’s Political Economy of AI Essay Collection, David Gray Widder and Mar Hicks draw on the history of tech hype cycles to warn against the harmful effects of the current generative AI bubble.
The year 2024 was dubbed “the largest election year in global history” with half the world’s population voting in national elections. Earlier this year, we hosted an event on AI and the 2024 Elections where scholars spoke about the potential influence of artificial intelligence on the election cycle– from misinformation to threats on election infrastructure. This webinar offered a reflection and exploration of the impacts of technology on the 2024 election landscape.
Earlier this year, the Allen Lab for Democracy Renovation hosted a convening on the Political Economy of AI. This collection of essays from leading scholars and experts raise critical questions surrounding power, governance, and democracy as they consider how technology can better serve the public interest.