Defining Public AI

A framework for public claims on resources

AI technologies promise to affect nearly every aspect of our lives.1 Knowing that this transformation will play out on a societal scale requires governance and ownership mechanisms that entitle a plurality of voices to steer AI not simply as a new category of tech products, but as a public resource and infrastructure.2 In democratic societies, the latter are subject to varying degrees of ‘publicness’ through (1) accessibility for use, (2) maintenance as a matter of public responsibility, and (3) accountability to the public in their function. Additionally, (4) participation in strategic decisions about the development and application of resources, and (5) how the value that is derived from these resources is distributed, which are typically harder to render public.3

Within complex innovation systems both public and private investments flow into research and development. Instead of understanding ‘public’ and ‘private’ as binary conditions attached to specific types of intermediaries (i.e. state versus market actors), FAE proposes to regard the notion of publicness as a spectrum on which the terms of public agency are negotiated, ranging from ‘thin’ to ‘thick’. ‘Thick’ public governance might apply to state-run assets funded by taxes, while ‘thin’ publicness could be obtained when/where infrastructures are freely accessible to the public but are privately owned. Between these extremes, frameworks such as commoning could be deployed to pool resources offering greater access and maintenance buy-in.4 5 Mission-driven public-private partnerships, such as those developed to build national supercomputing capabilities, leverage state-level accountability and strategic development while cooperating with private industry to deliver specialty products and services.6 7 Non-governmental interest organisations, academia, and grassroots movements also operate within this ecosystem to address unmet needs, both at the level of general policy and on behalf of specifically affected communities.8

Today’s AI is constituted through an entanglement of resources and infrastructures, each layer possessing its own context and openings for incorporating publicness into its design. What follows is the mapping of entanglements through AI’s technical stack, consisting of seven hierarchical layers organised in two tiers. The ‘hardware’ tier which provides the physical material and machinery by means of natural resource, server networks and compute layers. These enable the transfer and processing of information in the 'software', from the data layer, to the model, network protocols and application layers. By zooming in on the stack layers, the interdependence between industry, states, non-governmental organisations, academia, and the many publics that are implicated in the creation and adoption of AI can be understood. Highlighting the fact that ‘public AI’ is not just a speculative category, but a reality that requires ongoing development and support.9

Application Layer

Applications in the context of this AI stack are software products that utilise machine learning models as a core part of their capability; for example, content creation services such as Stable Diffusion, ChatGPT, or Suno.ai. Other present-day applications include virtual assistants (e.g. Apple’s Siri), recommendation systems (e.g. those used by Netflix), developer tools for writing code (e.g. GitHub’s CoPilot), speech and language recognition tools (e.g. Google Translate), biometric identification technology (e.g. fingerprint recognition AppLock), computer sensing and simulation systems (e.g. as those operated by driverless vehicles such as CARLA), search engines (e.g. Google Search and Bing), and others.

While the release of ChatGPT in November 2022 captured the public’s imagination, the history of applications that use different underlying AI capabilities (or precursors to AI) spans decades.10 Companies are, by default, incentivised to develop commercial products for consumers; however, governments, non-profits, and individuals also develop their own applications. For instance, in the UK the National Health Service develops applications for doctors to more accurately detect diseases using patient data.11 However, applications developed by the public sector are not always state-owned. Sometimes they are developed privately and licensed to the public sector. Regardless, privacy and consent are rights that are intrinsic to the individual irrespective of whether or not the tool is developed by the public sector, private sector, or state-owned.12

Application development is underpinned by a robust ecosystem of open research and code sharing. The development of applications relies on platforms such as GitHub (for code), and HuggingFace (for machine learning models and datasets). Applications can be patented (e.g. Spotify). In the US and Europe, applications are subject to a voluntary code of practice, which extends to application providers in general.13 14 Applications are also subject to a further round of scrutiny within digital marketplaces such as Apple’s App Store or Google Play.

Network Protocols Layer

Network protocols define the rules for how data is transmitted and received over a network, enabling AI applications to communicate efficiently with each other, with data sources, with AI models, and with end-users.

Hypertext Transfer Protocol (HTTP) is the open data communication protocol underlying the World Wide Web. Its specifications standardise the exchange of information; they are publicly available and can be used by anyone.15 As a protocol, it is not regulated by any single governmental entity. Instead, it is maintained and developed to international standards organisations, primarily the Internet Engineering Task Force (IETF) and the World Wide Web Consortium (W3C). These organisations work through consensus-driven processes involving various stakeholders, including developers, engineers, industry representatives, and others, to ensure the protocol remains effective, interoperable, and up to date with evolving technological needs.16

An Application Programming Interface (API), on the other hand, is a set of tools that enables the exchange of data and functionality between platforms, and integrations between different systems and devices. In AI applications, APIs are often used to access the capabilities of AI models, without interacting with the model itself, at a fee (e.g. enterprise-grade Gemini and ChatGPT Enterprise).

Model Layer

A machine learning model is a computer programme. In contrast to conventional programming, it is not manually defined through a sequence of instructions. Instead, the process of defining the computer programme is automated by means of algorithms that find patterns in large quantities of exemplary data.17 While there are many different machine learning methods, deep learning algorithms are the current predominant subset of methods. They typically use deep artificial neural network architectures with multiple layers that can recognise features and context in data.18 The model configuration is defined by a set of numerical parameters known as the model's weights. In 2021, the concept of a ‘foundation model’ was coined to describe a large (i.e. powerful) model that can process or generate information from multiple types of data inputs, such as text, images, audio, and video.19 Today these foundation models are largely multimodal, meaning they can move across those types of input and output. Since training large models from scratch is expensive, requiring large amounts of data and compute resources, foundation models provide an opportunity for commercialisation. Examples of commercial foundation models include OpenAI’s GPT-4 and Google’s Gemini, while Meta’s Llama 2 and the Mistral models are open source.20

Building efforts have consolidated around just a few foundation models within corporate, or privately-controlled, contexts due to the massive investments that they require.21 However, more recent projects have shown that smaller models are starting to compete with foundation ones.22 23 Nevertheless, only a handful of foundation models from US-based companies have the greatest number of users globally. There is debate as to whether foundation models should be required to be open source due to their impacts as a ‘foundational layer’ of most AI applications - Meta’s Llama 2, Stability AI’s Stable Diffusion and Google’s BERT are examples of open source foundation models emerging from proprietary companies - or whether the models should be subject to regulation by governments in jurisdictions where they are in use (or some combination of both).24 25

Whether open source or closed, not all model-makers reveal the contents of their training data, and it is believed that much of it is protected by copyright. Consequently, these leading AI organisations are confronting legal challenges that have yet to result in new laws.26 27 28 Meanwhile, the Responsible AI Licenses (RAIL) initiative is advocating for the development of licences for fine-tuning and downstream usage of models that curtails misuse.29

Data Layer

Models require high quality data for training and fine-tuning.30 While early efforts to create openly available, labelled datasets, for example ImageNet, improved the quality and capabilities of AI models, recent technical advances in foundation models have reduced the reliance on such resource-intensive efforts.31 32 Instead, the web is directly scraped by a large-scale web crawler, resulting in the Common Crawl dataset, which is a continually-updating set of raw webpage data with extracted metadata and text.33 Despite increasing criticism, it still serves as the most important resource for researchers, developers, and anyone interested in analysing the vast amount of information available on the internet.34

High profile lawsuits brought against companies including OpenAI, Microsoft, Stability AI, and Midjourney allege that their models infringe on copyrighted material in the training of their models.35 Even the use of openly accessible, or Creative Commons licensed material, has engendered significant debate around concerns relating to value extraction. Web crawling can also lead to datasets unintentionally containing illegal content. For example, the community-driven open source dataset LAION-5B has been accused of containing child abuse material.36 37 Other open data repositories, for example libraries on HuggingFace, offer developers easy access to specific datasets, while vendors such as Google Ads are given consent by users to own their data and then monetise it. Governments and the healthcare industry have high-quality datasets related to societies’ overall level of wellness through medical imaging and health records, as well as census and municipal data.38 39

Compute Layer

Data-driven machine learning algorithms are reliant on high-performance computing. Graphics processing units (GPUs) are semiconductor chips originally designed for 3D graphics, but their competency in performing complex mathematical calculations at high speeds has made them hardware that is fundamental to AI systems in order for models to be trained quickly and at scale.40 Nvidia, Advanced Micro Devices (AMD) and Intel are currently the largest companies producing GPU hardware. Additionally, organisations that offer applications that use computational power do not pay Nvidia directly for GPUs. Instead, they access GPUs via cloud providers such as Amazon Web Services, Microsoft Azure, and Google Cloud Platform, which, in turn, buy chips in bulk from companies including Nvidia.

Because of high demand, intense global competition and geopolitical tensions have emerged surrounding the semiconductor industry; nations are vying for dominance in manufacturing, design, and supply chain control. In the past three years, particularly in response to Taiwan’s geopolitical status as the leading global chip producer, governments in the US, the EU, and the UK have pushed to support national semiconductor manufacturing, research, and development through new policy positions, legislation, and by providing financial incentives. Generally, investment by outside actors in national or enterprise computing projects is regarded now as a matter of national security and global competition.41 Despite attempts by the US government to slow R&D capacity through restrictions on key exports in China and the Middle East, China is fast catching up to state-of-the-art chip manufacturing technologies. In addition to governments, industry players have also become active. OpenAI, for example, is seeking to raise $7tn, significantly from investors in the Arabian Gulf, for its own chip production capacity.42

As opposed to applications, network protocols, and small-scale models, the computing costs of foundational models means that GPUs are less accessible for small-scale entities or open innovation, raising questions about enclosure. It is a resource that has consolidated around a few key companies that service both the market and government needs.

Server Networks Layer

Server networks, otherwise known as ‘clouds’, are clusters of computers that store data, run software such as AI models, and provide access to both data and models via APIs and communication protocols.43 To operate on the internet, server networks rely on the vast system of underwater cables which act as highways for data traffic across the planet. Over recent years, major server networks like Amazon Web Services and Microsoft Azure have come to account for much of the web. The business model of server networks is a straightforward exchange of use/access for a fee. Companies that rely heavily on server architecture – such as Google – build their own.44 The strategic placement of servers near where they will be needed most has instigated a land grab by companies and governments as they try to secure space to build new server racks and cooling systems. National security is increasingly a major concern for countries as governments allow foreign companies to operate servers in their jurisdiction. In Guizhou, China, for instance, Apple operates the Chinese iCloud, a server network that is not connected to the global Apple iCloud.

Natural Resources Layer

Control over natural resources including oil, gas, and coal has shaped modern society, creating massive levels of wealth, and establishing new regulatory regimes, while simultaneously accelerating environmental breakdown and laying the foundation for the technological developments of the 20th and 21st centuries. In centuries past, the search for and sequestration of natural resources provided the foundation of colonial projects. Today, control over rare earth metals, integral to both the global rollout of renewable energy and the mass expansion of AI, will define relationships between countries and companies who have access to rare metal wealth and those who do not.

The natural resources required for AI can be broadly grouped under the headings of materials and energy. Materials e.g. silicon, gold, silver, palladium, and lithium are necessary for the fabrication of chips, servers, cables, and batteries. Energy from renewable sources (wind, solar, hydro), fossil fuels (oil, gas, coal), and nuclear reactions (fission and fusion) power the data centres, server networks, and computing operations, as well as their cooling systems.

From publicly owned oil companies such as Norway’s Equinor, to fully privatised water companies in the UK such as Thames Water, governance of natural resources across the Western world varies by location and resource. Much attention in recent years has been focused on establishing democratically, or at least more publicly, accountable energy systems across Europe with an accompanying shift away from oil, gas and coal towards renewables.45 This process has been accelerated by the massive profits generated by gas suppliers following the price spike resulting from the attack on Ukraine, and by accelerating climate breakdown. Less attention has been focused on the supply chains, and often weak governance models in markets for rare earth metals such as palladium and lithium, or even more traditional metal commodities, such as gold and silver. Often, these metals, key to a transition to renewable energy, are extracted from the territories of the former colonies of Europe, as transnational corporations accrue massive profits from this resource extraction at the expense of their workers and of local communities.46

Footnotes

  1. OpenAI, Developing safe & responsible AI [link].

  2. Shrey Jain et al., Plural Publics (2023) [link].

  3. Many groups are working to define ideas of public AI; they include: The Public AI Network, Public AI White Paper (2024) [link]; Collective Intelligence Project and The Alan Turing Institute amonst others. FAE4 draws on ideas from across these frameworks. We also use the term ‘public’ following John Dewey’s conception of the term in The Public and Its Problems (1927).

  4. This includes much of academia as well as non-profit orgs committed to stewarding the commons such as Internet Archive [link] and Arxiv [link].

  5. Elinor Ostrom, Governing the commons (1990).

  6. Mariana Mazzucato, Public Purpose: Industrial Policy’s Comeback and Government’s Role in Shared Prosperity, (2021).

  7. UK AI Research Resource dubbed Isambard-AI will be one of Europe’s most powerful supercomputers. The new facility will serve as a national resource for researchers and industry experts spearheading AI innovation and scientific discovery. An unprecedented £225m investment has been allotted to create UK's most powerful supercomputer in Bristol (2023) [link].

  8. These include organisations such as The Alan Turing Institute, Aapti Institute and Omydiar Network.

  9. An endeavour can be understood as ‘public’ when it is ‘in service of society and not industry or government’. See Jürgen Habermas, The Structural Transformation of the Public Sphere (1991).

  10. OpenAI, Introducing ChatGPT [link]; William van Melle, MYCIN: a knowledge-based consultation program for infectious disease diagnosis (1978) [link]; Bruce T. Lowerre, The HARPY Speech Recognition System (1976) [link]; Feng-hsiung Hsu, Behind Deep Blue: Building the Computer that Defeated the World Chess Champion (2002).

  11. Tammy Lovell, NHS rolls out AI tool which detects heart disease in 20 seconds (2022) [link].

  12. In France, tax authorities used proprietary software developed by Google to identify undeclared tax revenue. See* Undeclared pools in France uncovered by AI technology* [link].

  13. In 2022, the UK government set out a voluntary code of practice that includes better reporting of software vulnerabilities and more transparency for users regarding the privacy and security of apps available in all app stores. See New rules for apps to boost consumer security and privacy [link].

  14. In New York State, automated hiring apps are subject to bias testing. NYC Consumer and Worker Protection, See Automated Employment Decision Tools: Frequently Asked Questions (2023) [link].

  15. CERN, The birth of the Web [link].

  16. Wikipedia, HTTP [link].

  17. Rishi Bommasani et al., On the Opportunities and Risks of Foundation Models (2021) [link].

  18. Yann LeCun et. al, Deep Learning (2015) [link].

  19. Foundation models generally refer to transformer architecture trained on huge amounts of data and use transfer learning to perform general purpose tasks that can then be further fine-tuned into skills such as text synthesis, image manipulation, or audio generation. See Elliot Jones , Explainer: What is a foundation model? (2023) [link]. Rishi Bommasani et al.,_ On the Opportunities and Risks of Foundation Models _(2021) [link].

  20. See OpenAI’s GPT-4 [link]; Google’s Gemini [link]; Meta’s Llama 2 [link] and Mistral AI [link].

  21. Rishi Bommasani, et al., On the Opportunities and Risks of Foundation Models (2021) [link].

  22. Harsha Nori, Can Generalist Foundation Models Outcompete Special-Purpose Tuning? Case Study in Medicine (2023) [link].

  23. Yi Tay, Training Great LLMs Entirely from Ground Up in the Wilderness as a Startup (2024) [link].

  24. Billy Perrigo, Yann LeCun On How An Open Source Approach Could Shape AI (2024) [link].

  25. Stable Diffusion is a model by UK-based AI company Stability AI, with open source code and weights [link].

  26. See Tim Bradshaw and Joe Miller, New York Times sues Microsoft and OpenAI in copyright case (2023) [link].

  27. For example, an investigation by The Atlantic in August of 2023 revealed that Meta partially trained its extensive language model using a dataset named Books3, which includes over 170,000 books that are either unauthorised copies or otherwise protected under copyright rules. Alex Reisner, _Revealed: The Authors Whose Pirated Books Are Powering Generative AI _(2023) [link].

  28. Because their rise has been so meteoric, the attempts to regulate models are still in their infancy, with suggested measures under development in the EU, US, China, and the African Union. In the UK, the Frontier AI Taskforce, is a research team within the government to evaluate risks. In the House of Lords, a draft Artificial Intelligence (Regulation) Bill has been put forth. See Robert Hart, White House Unveils ‘Sweeping’ AI Strategy As Biden Pushes For Transparency And Safety (2023) [link] and Artificial Intelligence (Regulation) Bill (2023) [link].

  29. See Responsible AI Licenses (RAIL) initiative [link].

  30. High-quality data is accurate, complete, reliable, and relevant information for its intended use in operations, decision-making, analysis, or processing. See Maria Priestley et al., A Survey of Data Quality Requirements That Matter in ML Development Pipelines (2023). [link].

  31. Jia Deng et al., ImageNet: A Large-Scale Hierarchical Image Database (2009) [link].

  32. Richard Sutton, The Bitter Lesson (2019) [link].

  33. Common Crawl, Our Mission [link].

  34. Baack, Stefan, and Mozilla Insights, Training Data for the Price of a Sandwich: Common Crawl’s Impact on Generative AI (2024) [link].

  35. USA cases: Tremblay v OpenAI (consolidated with Silverman v OpenAI and Chabon v OpenAI), 2023; Alter v OpenAI and Microsoft (consolidated with Authors Guild & ors v OpenAI), 2023; Basbanes & Ngagoyeanes v Microsoft and OpenAI, 2024; The New York Times v Microsoft and OpenAI, 2023; Chabon & ors v Meta Platforms, Inc., 2023; Kadrey v Meta Platforms, Inc., 2023; Andersen v Stability AI, 2023; Getty Images v Stability AI, 2023; Huckabee & ors v Meta, Bloomberg, Microsoft and The EleutherAI Institute, 2023; J.Doe 1 and J.Doe 2 v GitHub, Microsoft and OpenAI, 2022; Concord Music Group & ors v Anthropic PBC, 2023; Thomson Reuters v Ross Intelligence, 2023. UK: Getty Images v Stability AI, 2023.

  36. LAION, Laion-5b: A New Era Of Open Large-Scale Multi-Modal Datasets (2022) [link].

  37. Alex J. Champandard on X (2023) [link].

  38. In the UK, Health Data Research UK is a portal that enables access to health data to enable research and development. See Health Data Research UK [link].

  39. Other public services, e.g. public transport or postal services, also create data that could be used to improve the services they provide. Over the last several years, ‘data dignity’ campaigners and associated organisations have been working to prototype new public governance models for data protection including data trusts and cooperatives. See RadicalxChange [link], Aapti Institute [link], Open Data Institute [link], Data Empowerment Fund [link], and Data Trusts Initiative [link].

  40. Significant increases in computational power since 2016, thanks to advancements by Nvidia, AMD, Intel, and Qualcomm, have enabled the training of larger and more complex AI systems [link].

  41. Paresh Dave, OpenAI Agreed to Buy $51 Million of AI Chips From a Startup Backed by CEO Sam Altman (2023) [link].

  42. Anna Tong et al., _Exclusive: ChatGPT-owner OpenAI is exploring making its own AI chips _(2023) [link].

  43. While it is true that server networks also use computing units (e.g. CPUs) to communicate, dividing the server network and compute layers in this AI tech stack allows for clarity in terms of the unique infrastructural and governance issues of each respective layer.

  44. Debbie Weinstein, Our $1 billion investment in a new UK data centre (2024) [link].

  45. TUC, Public ownership of clean power: lower bills, climate action, decent jobs (2022) [link].

  46. Jake Simms and Andy Whitmore, with contributions from Kim Pratt, Unearthing injustice: A global approach to transition minerals (2023) [link].