Panoramic: Automotive and Mobility 2025
The Federal Trade Commission (FTC) has re-opened and set aside its 2024 consent order against generative AI company, Rytr, signalling a shift in how the Commission will approach AI enforcement under President Trump's AI Action Plan and its mandate to remove barriers to AI innovation and leadership. This unusual step offers an early look at how the FTC may recalibrate enforcement involving AI products and services – particularly where products have both lawful and potentially unlawful uses – and underscores that the agency will require evidence of consumer harm before invoking its unfairness authority.
Last year, the FTC filed a complaint against Rytr, which currently markets itself as an “AI writing assistant for short-form content,” offering over 40 use cases ranging from emails and blog posts to social media ads and customer testimonials through both free and subscription tiers.
The FTC took particular issue with Rytr’s “Testimonial & Review” use case, alleging that the tool could “quickly generate an unlimited number of detailed and genuine-sounding reviews with minimal input,” and that some subscribers had generated tens of thousands of reviews –including in a single month. According to the FTC, the “Testimonial & Review” tool’s “likely only use” was to facilitate “fake reviews” to deceive customers and therefore had “no or de minimis legitimate use.” The FTC alleged two Section 5 violations for deception and unfairness and, in December 2024, Rytr entered a consent order (“Rytr Order”) that, among other obligations, banned it from offering or promoting any review or testimonial generation service.
Two Republican Commissioners dissented, laying the groundwork for the December 2025 set-aside.
One month later, President Trump issued Executive Order (EO) 14179, “Removing Barriers to American Leadership in Artificial Intelligence” and announced his administration would pursue an “AI Action Plan.” That plan, released earlier this year, requires the FTC to, among other things, “review all FTC final orders, consent decrees, and injunctions, and, where appropriate, seek to modify or set-aside any that unduly burden AI innovation.”
The two current members at the typically five-member Commission – Chair Andrew Ferguson and Commissioner Mark Meador, both Republicans – announced that the agency had reviewed the Rytr Order pursuant to the AI Action Plan and determined that the matter should be re-opened and set aside because it was against the public interest. They concluded that the Commission had not sufficiently pled any violation of Section 5 and that it unduly burdened AI innovation in violation of EO 14179 and the AI Action Plan. Rytr consented to vacating the order.
The set-aside focuses on the two theories of liability alleged in the initial matter: (1) “means and instrumentalities” of deception; and (2) unfairness.
The original complaint alleged that Rytr violated Section 5 of the FTC Act by providing its users and subscribers with the “means and instrumentalities to deceive” others by generating deceptive consumer reviews.
In vacating the Rytr Order, the current FTC leadership found the complaint did not allege conduct that met any of the traditional tests for “means and instrumentalities”-based liability. It did not allege Rytr made any deceptive statements; it did not allege the tool was inherently deceptive; and it did not allege Rytr knew or had reason to know that its tool was being used for deceptive purposes.
The Commission underscored that the mere potential for misuse does not establish a Section 5 violation. Echoing Chair Ferguson’s earlier dissent, the FTC warned that expanding the means and instrumentalities theory as applied in Rytr to tools with both lawful and unlawful uses could “condemn anyone who makes ‘pencils, paper, printers, computers, smartphones, word processors, typewriters, posterboard, televisions, billboards,’ or any other tool that someone somewhere might use to create or disseminate a false advertisement.”
The previous Commission alleged Rytr’s conduct met the Section 5 definition of unfairness primarily because the AI-generated customer review use case had “no or de minimis legitimate use.”
Current FTC leadership rejected that conclusion, finding that “almost all tools have both legal and illegal uses” and the very features that raised concern for the previous Commission – that is, the ease with which one could generate a customer review with minimal user effort – was, as then-Commissioner Holyoak described in her prior dissent, “a feature, not a bug.” Current FTC leadership also emphasized that an unfairness theory requires adequately alleging that the AI platform causes or is likely to cause substantial injury to consumers, and found that there was neither empirical nor qualitative evidence to support such a showing in the original Rytr Order.
Authored by Lance Murashige, Sophie Baum, Harsimar Dhanoa, and Soojin Jeong.