Download Executive Summary PDF

Download Full Report PDF


Artificial intelligence1The term ‘artificial intelligence’ has come to mean many different things over the course of its history, and may be best understood as a marketing term rather than a fixed object. See for example: Michael Atleson, “Keep your AI Claims in Check”, Federal Trade Commission, February 27, 2023, https://t.co/Z7np8sbIWs; Meredith Whittaker, “Signal, and the Tech Business Model Shaping Our World”, Conference on Steward-Ownership 2023, https://www.youtube.com/watch?v=blyDvc9dOEM,Annie Lowery, “AI Isn’t Omnipotent. It’s Janky”, The Atlantic, April 3, 2023, https://www.theatlantic.com/ideas/archive/2023/04/artificial-intelligence-government-amba-kak/673586/?utm_campaign=the-atlantic&utm_term=2023-04-03T11%3A30%3A57&utm_medium=social&utm_source=twitter&utm_content=edit-promo; is captivating our attention, generating both fear and awe about what’s coming next. As increasingly dire prognoses about AI’s future trajectory take center stage in the headlines about generative AI, it’s time for regulators, and the public, to ensure that there is nothing about artificial intelligence (and the industry that powers it) that we need to accept as given. This watershed moment must also swiftly give way to action: to galvanize the considerable energy that has already accumulated over several years towards developing meaningful checks on the trajectory of AI technologies. This must start with confronting the concentration of power in the tech industry.

The AI Now Institute was founded in 2017, and even within that short span we’ve witnessed similar hype cycles wax and wane: when we wrote the 2018 AI Now report, the proliferation of facial recognition systems already seemed well underway, until pushback from local communities pressured government officials to pass bans in cities across the United States and around the world.2Meredith Whittaker, Kate Crawford, Roel Dobbe, Genevieve Fried, Elizabeth Kaziunas, Varoon Mathur, Sarah Myers West, Rashida Richardson, Jason Schultz, Oscar Schwartz, AI Now 2018 Report, AI Now Institute, December 2018; Tom Simonite, “Face Recognition Is Being Banned—But It’s Still Everywhere,” Wired, December 22, 2021. Tech firms were associated with the pursuit of broadly beneficial innovation,3See Jenna Wortham, “Obama Brought Silicon Valley to Washington,” New York Times, October 25, 2016, https://www.nytimes.com/2016/10/30/magazine/barack-obama-brought-silicon-valley-to-washington-is-that-a-good-thing.html; and Cecilia Kang and Juliet Eilperin, “Why Silicon Valley Is the New Revolving Door for Obama Staffers,” Washington Post, February 28, 2015, https://www.washingtonpost.com/business/economy/as-obama-nears-close-of-his-tenure-commitment-to-silicon-valley-is-clear/2015/02/27/3bee8088-bc8e-11e4-bdfa-b8e8f594e6ee_story.html. until worker-led organizing, media investigations, and advocacy groups shed light on the many dimensions of tech-driven harm.4Varoon Mathur, Genevieve Fried, and Meredith Whittaker, “AI in 2019: A Year in Review,” Medium , October 9, 2019. 

These are only a handful of examples, and what they make clear is that there is nothing about artificial intelligence that is inevitable. Only once we stop seeing AI as synonymous with progress can we establish popular control over the trajectory of these technologies and meaningfully confront their serious social, economic, and political impacts—from exacerbating patterns of inequality in housing,5Robert Bartlett, Adair Morse, Richard Stanton, and Nancy Wallace, “Consumer-Lending Discrimination in the FinTech Era,” Journal of Financial Economics 143, no. 1 (January 1, 2022): 30–56. credit,6Christopher Gilliard. “Prepared Testimony and Statement for the Record,” Hearing on “Banking on Your Data: The Role of Big Data in Financial Services,” House Financial Services Committee Task Force on Financial Technology, 2019. healthcare,7Ziad Obermeyer, Brian Powers, Christine Vogeli, and Sendhil Mullainathan, “Dissecting Racial Bias in an Algorithm Used to Manage the Health of Populations,” Science 366, no. 6464 (October 25, 2019): 447–53. and education8Rashida Richardson and Marci Lerner Miller, “The Higher Education Industry Is Embracing Predatory and Discriminatory Student Data Practices,” Slate, January 13, 2021. to inhibiting workers’ ability to organize9 Ibid. and incentivizing content production that is deleterious to young people’s mental and physical health.10See Zach Praiss, “New Poll Shows Dangers of Social Media Design for Young Americans, Sparks Renewed Call for Tech Regulation,” Accountable Tech, March 29, 2023; and Tawnell D.  Hobbs, Rob Barry, and Yoree Koh, “‘The Corpse Bride Diet’: How TikTok Inundates Teens with Eating-Disorder Videos,” Wall Street Journal, December 17, 2021.

In 2021, several members of AI Now were asked to join the Federal Trade Commission (FTC) to advise the Chair’s office on artificial intelligence.11Federal Trade Commission, “FTC Chair Lina M. Khan Announces New Appointments in Agency Leadership Positions,” press release, November 19, 2021. This was, among other things, a recognition of the growing centrality of AI to digital markets and the need for regulators to pay close attention to potential harms to consumers and competition. Our experience within the US government helped clarify the path for the work ahead. 

ChatGPT was unveiled during the last month of our time at the FTC, unleashing a wave of AI hype that shows no signs of letting up. This underscored the importance of addressing AI’s role and impact, not as a philosophical futurist exercise but as something that is being used to shape the world around us here and now. We urgently need to be learning from the “move fast and break things” era of Big Tech; we can’t allow companies to use our lives, livelihoods, and institutions as testing grounds for novel technological approaches, experimenting in the wild to our detriment. Happily, we do not need to to draft policy from scratch: artificial intelligence, the companies that produce it, and the affordances required to develop these technologies already exist in a regulated space, and companies need to follow the laws we already in effect. This provides a foundation, but we’ll need to construct new tools and approaches, built on what we already have.

There is something different about this particular moment: it is primed for action. We have abundant research and reporting that clearly documents the problems with AI and the companies behind it. This means that more than ever before, we are prepared to move from identifying and diagnosing harms to taking action to remediate them. This will not be easy, but now is the moment for this work. This report is written with this task in mind: we are drawing from our experiences inside and outside government to outline an agenda for how we—as a group of individuals, communities, and institutions deeply concerned about the impact of AI unfolding around us—can meaningfully confront the core problem that AI presents, and one of the most difficult challenges of our time: the concentration of economic and political power in the hands of the tech industry—Big Tech in particular. 


There is no AI without Big Tech.

Over the past several decades, a handful of private actors have accrued power and resources that rival nation-states while developing and evangelizing artificial intelligence as critical social infrastructure. AI is being used to make decisions that shape the trajectory of our lives, from the deeply impactful, like what kind of job we get and how much we’re paid; whether we can access decent healthcare and a good education; to the very mundane, like the cost of goods on the grocery shelf and whether the route we take home will send us into traffic. 

Across all of these domains, the same problems show themselves: the technology doesn’t work as claimed, and it produces high rates of error or unfair and discriminatory results. But the visible problems are only the tip of the iceberg. The opacity of this technology means we may not be informed when AI is in use, or how it’s working. This ensures that we have little to no say about its impact on our lives. 

This is underscored by a core attribute of artificial intelligence: it is foundationally reliant on resources that are owned and controlled by only a handful of big tech firms. 

The dominance of Big Tech in artificial intelligence plays out along three key dimensions:

To understand how we got here, we need to look at how tech firms presented themselves in their incipiency: their rise was characterized by marketing rhetoric promising that commercial tech would serve the public interest, encoding democratic values like freedom, democracy, and progress. But what’s clear now is that the companies developing and deploying AI and related technologies are motivated by the same things that—structurally and necessarily—motivate all corporations: growth, profit, and rosy market valuations. This has been true from the start.


Why “Big Tech”?

In this report, we pay special attention to policy interventions that target large tech companies. The term “Big Tech” became popular around 201314Nick Dyer-Witheford and Alessandra Mularoni, “Framing Big Tech: News Media, Digital Capital and the Antitrust Movement,” Political Economy of Communication 9, no. 2 (2021): 2–20, https://polecom.org/index.php/polecom/article/view/145. as a way to describe a handful of US-based megacorporations, and while it doesn’t have a definite composition, today it’s typically used as shorthand for Google, Apple, Facebook, Amazon, and Microsoft (often abbreviated as GAFAM), and sometimes also includes companies like Uber or Twitter.  

It’s a term that draws attention to the unique scale at which these companies operate: the self-reinforcing network effects, data, and infrastructural advantages they have amassed enable them to box out competitors. Big Tech’s financial leverage has allowed these firms to consolidate this advantage across sectors from social media to healthcare to education and across media (like the recent pivot to virtual and augmented realities), often through strategic acquisitions. These firms seek to protect their advantage from regulatory threats through lobbying and similar non-capital strategies that leverage their deep pockets.15Zephyr Teachout and Lina Khan, “Market Structure and Political Law: A Taxonomy of Power,” Duke Journal of Constitutional Law & Public Policy 9, no. 1 (2014): 37–74. Following on from narratives around “Big Tobacco,” “Big Pharma,” and “Big Oil,” this framing draws on lessons from other domains where industrial consolidation of power and abusive practices led to movements to demand public accountability. (As one commentator puts it, “society does not prepend the label ‘Big’ with a capital B to an industry out of respect or admiration. It does so out of loathing and fear – and in preparation for battle.”16Will Oremus, “Big Tobacco. Big Pharma. Big Tech?” Slate, November 17, 2017.) Recent name changes, like Google to Alphabet or Facebook to Meta, also make the term Big Tech helpful in capturing the sprawl of these companies and their continually shifting contours.17Kean Birch Kean and Kelly Bronson, “Big Tech,” Science as Culture 31, no. 1 (January 2, 2022): 1–14.

Focusing on Big Tech is a useful prioritization exercise for tech policy interventions for several reasons:

Finally, this report takes aim not just at the pathologies associated with these companies, but also at the broader narratives that justify and normalize them. From unrestricted innovation as a social good to digitization, to data as the only way to see and interpret the world, to platformization as necessarily beneficial to society and synonymous with progress—and regulation as chilling this progress—these narratives pervade the tech industry (and, increasingly, government functioning as well). 


Strategic priorities

Where do we go from here? Across the chapters of this report, we offer a set of approaches that, in concert, will collectively enable us to confront the concentrated power of Big Tech. Some of these are bold policy reforms that offer bright-line rules and structural changes. Others aren’t in the traditional domain of policy at all, but acknowledge the importance of nonregulatory interventions such as collective action, worker organizing, while acknowledging the role public policy can play in bolstering or kneecapping these efforts. We also identify trendy policy responses that seem positive on their surface, but because they fail to meaningfully address power discrepancies should be abandoned. The primary jurisdictional focus for these recommendations is the US, although where relevant we point to policy windows or trends in other jurisdictions (such as the EU) with necessarily global impacts. 

Four strategic priorities emerge as particularly crucial to meet this moment: 

1. Place the burden on companies to affirmatively demonstrate that they are not doing harm, rather than on the public and regulators to continually investigate, identify, and find solutions for harms after they occur.

Investigative journalism and independent research has been critical to tech accountability: the hard work of those testing opaque systems has surfaced failures that have been crucial for establishing evidence for tech-enabled harms. But, as we outline in the section on Algorithmic Accountability, as a policy response, audits and similar accountability frameworks dependent on third-party evaluation play directly into the tech company playbook by positioning responsibility for identifying harms outside of the company. 

The finance sector offers a useful corollary for thinking this through. Much like AI, the actions taken by large financial firms have diffuse and unpredictable effects on the broader financial system and the economy at large. It’s hard to predict any particular harm these may cause, but we know the consequences can be severe, and the communities hit hardest are those that already experience significant inequality. After multiple crisis cycles, there’s now widespread consensus that the onus needs to be on companies to demonstrate that they are mitigating harms and to comply with regulations, rather than on the broader public to root these out. 

The tech sector, likewise has diffuse and unpredictable effects not only on our economy, but our information environment and labor market, among many other things. We see value in a due-diligence approach that requires firms to demonstrate their compliance with the law rather than turn to regulators or civil society to show where they haven’t complied—similar in orientation to how we already regulate many goods that have significant public impact, like food and medicine. And we need structural curbs like bright lines and no-go zones that identify types of use and domains of implementation that should be barred in any instance, as many cities have already established by passing bans on facial recognition. For example, in Algorithmic Management, we identify emotion recognition as a type of technology that should never be deployed, but particularly in the workplace: aside from the clear concerns about its use of pseudoscience and accompanying discriminatory effects, it is fundamentally unethical for employers to seek to draw inferences about their employees’ inner state to maximize their profit. And in Biometric Surveillance, we identify the absence of such bright-line measures as the animating force behind a slow creep of facial recognition and other surveillance systems into domains like cars and virtual reality. 

We also need to lean further toward scrutiny of harms before they happen rather than waiting to rectify harms after they’ve already occurred. We discuss what this might look like in the context of merger reviews in the Toxic Competition section, advocating for an approach to merger reviews that looks to predict and prevent abusive practices before they manifest, and in Antitrust, we break down how needed legal reforms would render certain kinds of mergers invalid in the first place, and put the onus on companies to demonstrate they aren’t anti-competitive.

2. Break down silos across policy areas, so we’re better prepared to address where advancement of one policy agenda impacts others. Firms play this isolation to their advantage.

One of the primary sources of Big Tech power is the expansiveness of their reach across markets, with digital ecosystems that stretch across vast swathes of the economy. This means that effective tech policy must be similarly expansive, attending to how measures adopted in the advancement of one policy agenda ramify across other policy domains. For example, as we underscore in the section on Toxic Competition, legitimate concerns about third-party data collection must be addressed in a way that doesn’t inadvertently enable further concentration of power in the hands of Big Tech firms. Disconnection between the legal and policy approaches to privacy on the one hand and competition on the other have enabled firms to put forward self-regulatory measures like Google’s Privacy Sandbox in the name of privacy that ultimately will lead to the depletion of both privacy and competition by strengthening Google’s ability to collect information on consumers directly while hollowing out its competitors. These disconnects can also prevent progress in one policy domain from carrying over to another. Despite years of carefully accumulated evidence on the fallibility of AI-based content filtration tools, we’re seeing variants of the magical thinking that AI tools will be able to scan effectively, even perfectly, for illegal content, crop up once again in encryption policy with the EU’s recent “chat control” client-side scanning proposals.18Ross Anderson, “Chat Control or Child Protection?”, University of Cambridge Computer Lab, October 13, 2022 https://www.cl.cam.ac.uk/~rja14/Papers/chatcontrol.pdf

Policy and advocacy silos can also blunt strategic creativity in ways that foreclose alliance or cross-pollination. We’ve made progress on this front in other domains, ensuring for example that privacy and national security are increasingly seen as consonant, rather than mutually exclusive, objectives. But AI policy has been undermined too often by a failure to understand AI materially, as a composite of  data, algorithmic models, and large-scale computational power. Once we view AI this way, we can understand data minimization and other approaches that limit data collection not only as protecting consumer privacy, but as mechanisms that help mitigate some of the most egregious AI applications, by reducing firms’ data advantage as a key source of their power and rendering certain types of systems impossible to build.  It was through data protection law that Italy’s privacy regulator was the first to issue a ban on ChatGPT19Clothilde Goujard, “Italian Privacy Regulator Bans ChatGPTPolitico, March 31, 2023. and, the week before that, Amsterdam’s Court of Appeal ruled automated firing and opaque algorithmic wages to be illegal.20Worker Info Exchange, “Historic Digital Rights Win for WIE and the ADCU Over Uber and Ola at the Amsterdam Court of Appeals“, April 4, 2023 FTC officials also recently called for leveraging antitrust as a tool to enhance worker power, including to push back against worker surveillance.21Elizabeth Wilkins, “Rethinking Antitrust“, March 30, 2023 This opens up space for advocates working on AI-related issues to form strategic coalitions with those that have been leveraging these policy tools in other domains.  This multivariate approach has the added advantage of necessitating that those focused on AI-related issues form strategic coalitions with those that have been leveraging these policy tools in other domains. 

Throughout this report, we attempt to establish links between related, but often siloed domains. Namely, we look at the ways that data protection and competition reform can act as AI policy (see section on Data Minimization; Antitrust), and we look at the way AI policy often doubles as industrial policy (see section on Algorithmic Accountability). 

3. Identify when policy approaches get co-opted and hollowed out by industry, and pivot our strategies accordingly. 

The tech industry, with its billions of dollars and deep political networks, has been both nimble and creative in its response to anything perceived as a policy threat. There are relevant lessons from the European experience around the perils of shifting from a “rights-based” regulatory framework, as in the GDPR, to a “risk-based” approach, as in the upcoming AI Act and how the framing of “risk” (as opposed to rights)  could tip the playing field in favor of industry-led voluntary frameworks and technical standards.22Fanny Hidvegi and Daniel Leufer, “The EU should regulate AI on the basis of rights, not risks”, Access Now, February 17, 2021, https://www.accessnow.org/eu-regulation-ai-risk-based-approach/

Responding to the growing chorus calling for bans on facial recognition technologies in sensitive social domains, several tech companies pivoted from resisting regulation to claiming to support it, something they often highlighted in their marketing. The fine print showed that what these companies actually supported were soft moves positioned to undercut bolder reform. For example, Washington state’s widely critiqued facial recognition law passed with Microsoft’s support. The bill prescribed audits and stakeholder engagement, a significantly weaker stance than banning police use which is what many advocates were calling for (see section on Biometric Surveillance).

For example, mountains of research and advocacy demonstrate the discriminatory impacts of AI systems and the fact that these issues cannot be addressed solely at the level of code and data. While the AI industry has accepted that bias and discrimination is an issue, companies have also been quick to narrowly cast bias as a technical problem with a technical fix. 

Civil society responses must be nimble in responding to Big Tech subterfuge, and we must learn to recognize such subterfuge early. We draw from these lessons when we argue that there is disproportionate policy energy being directed toward AI and algorithmic audits, impact assessments, and “access to data” mandates. Indeed, such approaches have the potential to eclipse and nullify structural approaches to curbing the harms of AI systems (see section on Algorithmic Accountability). In an ideal world, such transparency-oriented measures would live alongside clear standards of accountability and bright-line prohibitions. But this is not what we see happening. Instead, a steady stream of proposals position algorithmic auditing as the primary policy approach toward AI. 

Finally, we also need to stay on top of companies’ moves to evade regulatory scrutiny entirely: for example, firms have been seeking to introduce measures in global trade agreements (see section on Global Digital Trade) that would render regulatory efforts seeking accountability by signatory countries presumptively illegal. And companies have sought to use promises of AI magic as a means of evading stronger regulatory measures, such as by clinging to the familiar false argument that AI can provide a fix for unsolvable problems, such as in content moderation.23Federal Trade Commission, “Combatting Online Harms Through Innovation”, Federal Trade Commission, June 2022.

4. Move beyond a narrow focus on legislative and policy levers and embrace a broad-based theory of change.

To make progress and ensure the longevity of our wins, we must be prepared for the long game, and author strategies that keep momentum going in the face of inevitable political stalemates. We can learn from ongoing organizing in other domains, from climate advocacy (see section on Climate) that identifies the long-term nature of these stakes, to worker-led organizing (see section on Algorithmic Management) which has emerged as one of the most effective approaches to challenging and changing tech company practice and policy. We can also learn from shareholder advocacy (see section on Tech & Financial Capital), which uses companies’ own capital strategies to push for accountability measures – one example is the work of the Sisters of St. Joseph of Peace using shareholder proposals to hold Microsoft to account for human rights abuses. The Sisters also used such proposals to seek a ban on the sale of facial recognition to government entities, and to require Microsoft to evaluate how the company’s lobbying aligns with its stated principles.24See Chris Mills Rodrigo, “Exclusive: Scrutiny Mounts on Microsoft’s Surveillance Technology,” Hill, June 17, 2021, https://thehill.com/policy/technology/558890-exclusive-scrutiny-mounts-on-microsofts-surveillance-technology; and Issie Lapowsky, “These Nuns Could Force Microsoft to Put Its Money Where Its Mouth Is,” Protocol, November 19, 2021, https://www.protocol.com/policy/microsoft-lobbying-shareholder-proposal. Across these fronts, there is much to learn from the work of organizers and advocates well-versed in confronting corporate power. 


Windows for action: The AI policy landscape

These strategic priorities are designed to take advantage of current windows for action. We summarize them below, and review each in more detail in the body of the report.  

1. Unwind tech firms’ data advantage.


2. Reform competition law and enforcement such that they can more capably reduce tech industry concentration.


3. Regulate ChatGPT, BARD, and other large-scale models.


4. Displace audits as the primary policy response to harmful AI.


5. Future-proof against the quiet expansion of biometric surveillance into new domains like cars.


6. Enact strong curbs on worker surveillance.


7. Prevent “international preemption” by digital trade agreements that can be used to weaken national regulation on algorithmic accountability and competition policy.


It’s time to move: years of critical work and organizing has outlined a clear diagnosis of the problems we face, regulators are primed for action, and we have strategies ready to be deployed immediately for this effort. We’ll also need more: those engaged in this work are out-resourced and out-flanked amidst a significant uptick in industry lobbying and a growing attack on critical work, from companies firing AI Ethics teams to universities shutting down critical research centers. And we face a hostile narrative landscape. The surge in AI hype that opened 2023 has moved things backwards, re-introducing the notion that AI is associated with ‘innovation’ and ‘progress’ and drawing considerable energy toward far-off hypotheticals and away from the task at hand.

We intend this report to provide strategic guidance to inform the work ahead of us, taking a bird’s eye view of the landscape and of the many levers we can use to shape the future trajectory of AI – and the tech industry behind it – to ensure that it is the public, not industry, that this technology serves – if we let it serve at all.