FTC Gears Up for AI Enforcement: No Brakes in Sight
FTC Gears Up for AI Enforcement: No Brakes in Sight
It is full speed ahead for the Federal Trade Commission’s continuing efforts to understand and determine how to regulate AI. On January 25, 2024, the FTC held its inaugural AI summit.[1] In alignment with the agency’s first algorithmic-bias enforcement action in December and its recent announcement of 6(b) Orders investigating industry’s relationships with and investments in generative AI companies, the summit focused on the threat AI could pose to consumer privacy and the impact of market concentration on competition and innovation. The FTC, eager to avoid what it perceives as its own mistakes concerning corporate consolidations of the 1990s and 2000s,[2] previewed its enforcement priorities for the coming year.
To start the summit, FTC Chair Lina Khan announced the FTC has opened an inquiry into generative AI partnerships and investments. Foreshadowing potential enforcement actions, the Commission issued 6(b) Orders to three large cloud service providers and two major AI software companies.[3] The move is significant because the FTC is examining partial ownership and partnership arrangements, a sub‑category of mergers recently highlighted in the FTC’s December 2023 Merger Guidelines jointly released with the DOJ Antitrust Division.
The announcement follows the FTC’s November 23, 2023, omnibus resolution authorizing its staff to issue compulsory process requests for non-public investigations involving products and services that use or claim to be produced by AI. The resolution helps to streamline the FTC’s investigations.[4]
The summit consisted of three panels addressing: (1) the role that chips and cloud providers have in AI; (2) the role of data in AI technologies and models; and (3) AI in consumer applications. Panelists articulated concerns about the robustness of competition throughout the AI tech stack and ethical AI deployment. Some of the concerns discussed included:
According to Commissioner Rebecca Kelly Slaughter, the concentration of AI models with access to consumer data in the hands of a few companies could pose risks around decisions and access to opportunity, information, and privacy. Panelists collectively advocated for the importance of consumer protection in AI, focusing on the implications for privacy and data usage. And both FTC and Consumer Financial Protection Bureau representatives repeatedly noted that “there is no AI exception to the law.”
In his remarks, Commissioner Bedoya highlighted that both old and new AI or algorithmic tools can cause potential violations, such as unfair discrimination.[5] According to him, AI technology success looks like:
The FTC views claims made about AI in marketing materials as a potential enforcement frontier. False advertising claims are of particular interest to the FTC, as such advertisements were characterized by Commissioner Slaughter as both anti-competitive and harmful to consumers. The FTC might also focus on claims around how data is used in connection with AI.
Samuel Levine, director of the FTC’s Bureau of Consumer Protection, closed the day by detailing the bureau’s efforts, including issuing guidance, pursuing enforcement actions, and proposing rules to address AI challenges. He summarized what the public can expect going forward:
Laura Miller contributed to authoring this client alert.
[1] A transcript of the summit is now available online.
[2] AI Tech Summit, Remarks by Chair Lina Khan, Jan. 25, 2024.
[3] FTC Launches Inquiry into Generative AI Investments and Partnerships, Jan. 25, 2024
[4] FTC Authorizes Compulsory Process for AI-related Products and Services, Nov. 21, 2023
[5] In the past decade, the FTC has repeatedly raised concerns about discrimination generated from data analytics. For example, in 2016, the FTC warned companies that the use of non-representative datasets as inputs may lead algorithms to produce biased results. See Federal Trade Commission, Big Data: A Tool for Inclusion or Exclusion? Understanding the Issues (FTC Report), January 2016, p. 28.
[6] The FTC recently required an AI model to be deleted in its enforcement action against Rite Aid for using racially biased facial recognition security surveillance tools. Prior to Rite Aid, the FTC mandated AI model deletion in several cases related to models or algorithms developed from certain data. See Final Order, In the Matter of Cambridge Analytica, LLC, FTC File No. 18323107 (Dec. 6, 2019); and U.S. v. Kurbo, Inc. et al., FTC Matter No. 1923228 (Mar. 4, 2022).
Practices
Industries + Issues