About
Consumer Privacy Class Action

Consumer Privacy Class Action

Tracking class action litigation alleging consumer privacy violations, data breaches, and unauthorized data collection.

13 entries in Legal Intelligence Tracker

Fashion, Beauty, Wearable Brands Face Stricter 2026 Privacy Rules

Fashion, beauty, and wearable technology companies face a fundamentally reshaped data privacy regime in 2026. New omnibus consumer privacy laws in California, Connecticut, Indiana, Kentucky, Rhode Island, Washington, and Nevada—combined with the EU's AI Act and heightened FTC enforcement—have elevated privacy from a compliance checkbox to a core product and marketing consideration. The shift is driven by three specific regulatory pressures: biometric data (facial mapping and body scanning in virtual try-on tools) now classified as sensitive personal information; consumer health data from wearables tracking stress, sleep, and menstrual cycles, regulated outside HIPAA by states including Connecticut and Washington; and strengthened children's privacy protections through state laws and California's Age-Appropriate Design Code. Class-action litigants are simultaneously challenging tracking and cookie practices under state wiretap statutes like California's CIPA.

Florida AG Investigates OpenAI, ChatGPT, Citing National Security Risks, FSU Shooting

Florida Attorney General James Uthmeier announced on April 9, 2026, that his office is launching an investigation into OpenAI and its ChatGPT models, alleging their role in facilitating a 2025 Florida State University (FSU) shooting, harming minors, enabling criminal activity, and posing national security risks from potential exploitation by adversaries like the Chinese Communist Party.[1][2][3][4][5][6][7] Subpoenas are forthcoming, with probes focusing on ChatGPT's alleged assistance to the FSU gunman—who queried it on the day of the April 17, 2025, attack about public reaction to a shooting and peak times at the FSU student union—plus links to child sex abuse material, grooming, and suicide encouragement.[1][3][5][6][7]

Florida court tosses DPPA parking citation lawsuit over lack of injury

A federal judge in the Southern District of Florida dismissed a class-action lawsuit under the Driver's Privacy Protection Act against Professional Parking Management Corporation, finding the plaintiff lacked Article III standing. The suit alleged the company used license plate readers in private parking lots, cross-referenced plates against state DMV records without consent, and mailed notices demanding $94.99—styled to resemble official citations—for unpaid parking charges. The plaintiff sought nationwide class certification and added Florida consumer-protection claims.

FTC and Congress intensify surveillance pricing crackdown amid state legislative wave

Federal regulators and lawmakers are moving aggressively against surveillance pricing—the practice of using consumer data to set individualized prices for identical products or services. In April 2026, FTC leadership told Congress that staff work on the issue continues, with the agency considering whether new disclosure requirements should apply to highly personalized, data-driven pricing. That same month, the House Oversight Committee launched a formal investigation, sending letters to major travel and platform companies demanding documentation on revenue management algorithms, consumer data practices, and testing protocols.

Washington Gov. Ferguson Signs HB 2225 Requiring AI Companion Chatbot Disclosures

Washington State Governor Bob Ferguson signed House Bill 2225, the Chatbot Disclosure Act, into law on March 24, 2026, effective January 1, 2027. The statute requires operators of "companion" AI chatbots—systems designed to simulate human responses and sustain ongoing user relationships—to disclose at the outset of interactions and every three hours (hourly for minors) that the bot is artificially generated. The law prohibits chatbots from claiming to be human, mandates protocols for detecting self-harm or suicidal ideation, bans manipulative engagement tactics targeting minors such as encouraging secrecy from parents or prolonged use, and bars sexually explicit content for underage users. Exemptions carve out business operational bots, gaming features outside sensitive topics, voice command devices, and curriculum-focused educational tools. Violations constitute unfair or deceptive acts under the Washington Consumer Protection Act (RCW 19.86), enforceable by the Attorney General and through private right of action allowing consumers to recover actual damages up to $25,000 treble.

Tesla Owners Sue Over Unfulfilled FSD Promises on HW3 Hardware

Tesla faces coordinated class-action litigation across multiple jurisdictions from owners of Hardware 3-equipped vehicles manufactured between 2016 and 2024. The plaintiffs allege that Tesla and Elon Musk made false representations that these vehicles would achieve full self-driving capability through software updates alone. A spring 2026 software release exposed Hardware 3's technical limitations, effectively excluding millions of owners from advanced autonomous features now reserved for newer Hardware 4 systems. The lead case, brought by retired attorney Tom LoSavio, centers on buyers who paid $8,000 to $12,000 for full self-driving capability that is now incompatible with their vehicles without costly hardware retrofits Tesla has not formally offered. Similar suits have been filed in Australia, the Netherlands, across Europe, and in California, where one action involves approximately 3,000 plaintiffs. Globally, the disputes affect roughly 4 million vehicles.

CT AG Tong Issues Feb. 25 Memo Applying Existing Laws to AI

Connecticut Attorney General William Tong issued a memorandum on February 25, 2026, clarifying how existing state law applies to artificial intelligence systems. The advisory targets four enforcement areas: civil rights laws prohibiting AI-driven discrimination in hiring, housing, lending, insurance, and healthcare; the Connecticut Data Privacy Act, which requires companies to disclose AI use, obtain consent for sensitive data collection, minimize data retention, conduct protection assessments for high-risk AI processing, and honor consumer deletion rights even within trained models; data safeguards and breach notification requirements; and the Connecticut Unfair Trade Practices Act and antitrust laws, which address deceptive AI claims, fake reviews, robocalls, and algorithmic price-fixing. The memorandum applies broadly to any business deploying AI in consequential decisions and specifically references harms including AI-generated nonconsensual imagery on platforms like xAI's Grok.

Workers File 7 Class-Action Lawsuits Against Mercor Over Data Breach Exposure[1][2]

Mercor, a $10 billion San Francisco AI startup that supplies training data to OpenAI, Anthropic, and Meta, is defending itself against at least seven class-action lawsuits filed in recent weeks. The suits stem from a data breach last month that exposed contractor information including recorded job interviews, facial biometric data, computer screenshots, and background checks. Plaintiffs allege Mercor violated federal privacy regulations by collecting extensive data through monitoring software like Insightful, sharing it with AI partners, and using interviews and proprietary materials to train models without adequate consent or disclosure.

FTC Reports $2.1B Losses from Social Media Scams in 2025

The Federal Trade Commission released data on April 27, 2026, documenting $2.1 billion in reported losses from social media scams during 2025—making them the costliest fraud contact method on record. Nearly 30 percent of victims who lost money reported the fraud originated on social media, an eightfold increase from 2020. Facebook accounted for the largest share of losses, exceeding WhatsApp and Instagram combined and surpassing text or email scams individually.

Three New State Privacy Laws Activate January 1, 2026, Expanding U.S. Patchwork to 20 States

Three new comprehensive consumer privacy laws took effect on January 1, 2026, in Indiana, Kentucky, and Rhode Island, bringing the total number of active state privacy regimes to 20. These laws grant consumers rights to access, correct, delete, and port their data, require opt-in consent for sensitive data processing, and impose civil penalties ranging from $7,500 to $10,000 per violation, enforced by state attorneys general. Simultaneously, California's DELETE Act (SB 362) will operationalize a centralized data broker deletion platform by August 1, 2026, with $200 daily fines per unfulfilled request beginning January 31. The CCPA has also been amended to require cybersecurity audits, risk assessments, and automated decision-making disclosures.

Ninth Circuit Affirms Dismissal of Brita Filter Class Action on April 16, 2026[1][2][6]

On April 16, 2026, the Ninth Circuit affirmed dismissal of a consumer class action against Brita Products Company, holding that a reasonable consumer would not expect a $15 water filter to remove all hazardous contaminants. Plaintiff Nicholas Brown sued under California's Unfair Competition Law, False Advertising Law, and Consumers Legal Remedies Act, claiming Brita's labels for its Everyday Pitcher and Standard Filter misled buyers into believing the products eliminated contaminants like arsenic, chromium-6, PFOA, PFOS, nitrates, and radium to undetectable levels. The three-judge panel, led by Judge Kim McLane Wardlaw, rejected the claims after the Los Angeles district court had already dismissed without leave to amend in September 2024.

7th Circuit Rules 2024 BIPA Damages Amendment Applies Retroactively to Pending Cases

On April 1, 2026, the U.S. Court of Appeals for the Seventh Circuit unanimously held that Illinois' August 2024 amendment to the Biometric Information Privacy Act applies retroactively to all pending cases. In Clay v. Union Pacific Railroad Co. (consolidated with Willis and Gregg), the court classified the amendment as procedural rather than substantive, allowing it to govern cases filed before its effective date. The amendment fundamentally restructures BIPA damages by capping recovery at $1,000 per violation for negligent violations and $5,000 for intentional ones—eliminating the "per-scan" theory that previously allowed plaintiffs to multiply damages across each biometric collection or transmission event.

LawSnap Briefing Updated May 11, 2026

State of play.

  • Pixel-based and tracking-tool privacy litigation is fragmenting across circuits and statutes. The Second Circuit has reinforced its "ordinary person" test to dismiss VPPA pixel claims against NBCUniversal, district courts in California are expanding CCPA's private right of action to cover tracking disclosures without a traditional breach, and the Third Circuit is redirecting state wiretap claims back to state courts for lack of Article III standing (→ Second Circuit Affirms Dismissal of VPPA Class Action Against NBCUniversal[1][3]).
  • DPPA standing doctrine is tightening in federal court. The Southern District of Florida dismissed Cicale v. Professional Parking Management Corporation with prejudice—finding that receiving a collections notice and paying a legitimate debt does not constitute concrete injury—while parallel DPPA cases involving Carfax's crash-report data in Maryland continue surviving dismissal, confirming that standing, not the merits, is now the dispositive battleground (→ Florida court tosses DPPA parking citation lawsuit over lack of injury).
  • Biometric and wearable health data exposure is accelerating across the fashion, beauty, and wearable tech sectors, with virtual try-on tools, wearable health monitors, and cookie-based tracking practices now drawing simultaneous CIPA class action filings and state AG scrutiny under a reshaped 2026 multi-state privacy regime (→ Fashion, Beauty, Wearable Brands Face Stricter 2026 Privacy Rules).
  • Virginia is poised to end its 175-year ban on state-court class actions, with legislation effective January 1, 2027 eliminating the consumer reliance requirement under the VCPA and opening a new state-court forum for statutory damages claims against consumer-facing businesses (→ Virginia Poised to Sign Class Action Law, Ending 175-Year Ban).
  • For counsel advising companies with consumer-facing digital products, subscription services, data collection operations, or edtech vendor relationships, the practical baseline is simultaneous exposure across five vectors—state privacy enforcement, federal ROSCA/FTC Act claims, a circuit-dependent tracking-claim pleading landscape, a newly activated Virginia state-court forum opening in 2027, and a rapidly escalating breach-litigation environment where vendor access controls and API security are now pleading targets.

Where things stand.

  • VPPA pixel litigation has a firm circuit split. The Second Circuit applies an "ordinary person" test that has dismissed multiple pixel-based VPPA claims including the NBCUniversal action; the First Circuit has taken different approaches, making venue selection material for both plaintiffs and defendants (→ Second Circuit Affirms Dismissal of VPPA Class Action Against NBCUniversal[1][3]).
  • CCPA's private right of action is expanding beyond breach. District court rulings in Shah v. Capital One and a Therapymatch case have allowed CCPA claims to proceed based on unauthorized disclosure through tracking tools to third parties—no traditional breach required—departing sharply from earlier precedent .
  • Article III standing doctrine is actively sorting tracking claims by data sensitivity. Courts allow pixel-tracking claims to survive when sensitive health data is exposed, while dismissing claims based on routine behavioral data without sensitive information attached; the DPPA standing dismissal in Cicale reinforces that tangible injury beyond data misuse is required across privacy statutes (→ Florida court tosses DPPA parking citation lawsuit over lack of injury).
  • Cookie banner compliance has become an independent litigation vector. CIPA claims targeting non-functional "Reject All" buttons and dark-pattern consent interfaces are proliferating; Honda and HelloFresh have already resolved enforcement actions, and over 1,000 CIPA suits were filed in 2025 alone .
  • ROSCA enforcement against subscription dark patterns is active and expanding. The FTC's Uber case survived dismissal on the core theory that pre-stored payment credentials cannot substitute for fresh affirmative consent before subscription enrollment; 21 state AGs are co-plaintiffs .
  • Junk fee class actions and mass arbitrations are accelerating. The FTC's Rule on Unfair or Deceptive Fees is in force for live-event tickets and short-term lodging; California's SB 478 adds per-violation penalties; plaintiffs' firms are bypassing class-action waivers through coordinated mass arbitrations (→ Surge in "Junk Fee" Class Actions Targets Hidden Pricing Practices).
  • California CPPA is enforcing opt-out fragmentation. The agency's 2026 enforcement actions target businesses that honor opt-outs in some contexts but not others—fragmented compliance is itself the violation .
  • State AG enforcement is active across consumer protection vectors. Coordinated multi-state actions target deceptive pricing in retail, rental housing junk fees, and financial services; state AGs are co-plaintiffs in the Uber ROSCA action; the Louisiana AG secured a $45 million settlement with CVS Health over deceptive practices (→ Federal and State Regulators Target Grocery Chains, Landlords, MLMs, and Credit Agencies).
  • The Ninth Circuit's reasonable consumer standard is being refined in both directions. The court dismissed the Brita filter action—price point and qualified language ("reduces") defeated the claim—while reviving the Target thread count action on the theory that literal falsity on an objective specification cannot be defeated by an ambiguity defense the defendant hasn't established (→ Ninth Circuit Affirms Dismissal of Brita Filter Class Action on April 16, 2026[1][2][6], Ninth Circuit Revives Target Thread Count Class Action[1][7]).
  • State privacy law proliferation continues without federal resolution. Alabama enacted the 21st comprehensive state privacy statute; the SECURE Data Act has been introduced in the House with full state-law preemption but no Democratic support; Indiana, Kentucky, and Rhode Island privacy laws took effect January 1, 2026 .

Latest developments.

Active questions and open splits.

  • CCPA private right of action scope: breach-only or tracking-disclosure? The Shah and Therapymatch rulings extend CCPA liability to third-party tracking disclosures without a breach—a significant departure from prior precedent that has not yet been tested at the appellate level. Whether the Ninth Circuit endorses this expansion will determine class action exposure for the entire California-facing digital economy .
  • VPPA "ordinary person" test: circuit divergence is now actionable. The Second Circuit's NBCUniversal ruling hardens the defendant-favorable standard while other circuits remain more plaintiff-permissive. Defendants in the Second Circuit have strong grounds for dismissal; plaintiffs are likely to forum-shop toward the First Circuit and others (→ Second Circuit Affirms Dismissal of VPPA Class Action Against NBCUniversal[1][3]).
  • DPPA standing: does the data-commercialization model determine survival? Cicale dismissed a parking enforcement DPPA claim for lack of injury while the Carfax crash-report case in Maryland survived—suggesting courts are distinguishing between incidental DMV data use and systematic commercial exploitation. The line between those models is not yet defined by any circuit court (→ Florida court tosses DPPA parking citation lawsuit over lack of injury).
  • Biometric and wearable health data: which regulatory regime governs? Consumer health data from wearables tracking stress, sleep, and menstrual cycles falls outside HIPAA but within state health data statutes in Connecticut and Washington—and potentially within biometric data frameworks in Illinois and other states. The classification question determines consent obligations, breach exposure, and which private right of action applies (→ Fashion, Beauty, Wearable Brands Face Stricter 2026 Privacy Rules).
  • Article III standing for tracking claims: sensitive vs. non-sensitive data. Courts are drawing a line between health-related data disclosures—which can constitute injury-in-fact without financial harm—and routine behavioral data, which cannot. The pleading distinction is now central to survival at the motion to dismiss stage, and the Cicale DPPA dismissal reinforces that paying a legitimate underlying obligation forecloses financial harm theories entirely (→ Florida court tosses DPPA parking citation lawsuit over lack of injury).
  • Federal preemption: will the SECURE Data Act displace state privacy regimes? The bill's preemption language would eliminate the CCPA, Virginia CDPA, and 19 other state frameworks if enacted—but it lacks Democratic support and faces a long history of failed federal privacy efforts. The preemption question is the central advisory issue for multistate compliance programs .
  • Canvas breach: what vendor-oversight duties do institutional clients bear? The recompromise of Instructure's systems after initial containment—exploited through Free-for-Teacher API access—raises unresolved questions about whether schools and universities face independent exposure under state privacy laws or their contractual obligations to Canvas users, and what due diligence standard applies to edtech vendor selection and monitoring .

What to watch.

  • Whether any circuit court takes up the CCPA tracking-disclosure expansion on appeal, and whether California's CPPA files its own enforcement action on the same theory .
  • Whether a circuit court addresses the DPPA data-commercialization split—distinguishing parking enforcement from Carfax-style crash-report sales—and whether that line becomes the organizing principle for future DPPA standing analysis (→ Florida court tosses DPPA parking citation lawsuit over lack of injury).
  • Whether Governor Spanberger signs Virginia's class action legislation and whether early litigation tests the venue restrictions and the broadened VCPA standard before the January 2027 effective date (→ Virginia Poised to Sign Class Action Law, Ending 175-Year Ban).
  • Whether class action filings in the Canvas breach consolidate into MDL proceedings and whether courts treat Instructure's API access controls and Free-for-Teacher account architecture as independent negligence and state privacy law pleading targets .
  • SECURE Data Act committee markup and whether preemption language survives or is narrowed in response to Democratic opposition and state AG pushback .
  • State AG enforcement actions targeting cookie and pixel-tracking practices in the fashion, beauty, and wearable tech sectors—the first enforcement wave under the 2026 multi-state privacy regime will set the compliance baseline for biometric and health data handling (→ Fashion, Beauty, Wearable Brands Face Stricter 2026 Privacy Rules).

mail Subscribe to Consumer Privacy Class Action email updates

Primary sources. No fluff. Straight to your inbox.

Also on LawSnap