ACCEPTING CASES NATIONWIDE

585.287.8188

CALL US

CONTACT US

Written By: Robert King, Esq.
Legal Review By: Jerry King, Esq.
The Grok Lawsuit is an Active Lawsuit
See if You Qualify

The Grok lawsuit involves legal claims against xAI, which people used to make sexualized deepfakes, often involving women and children. The lawsuits allege that the company’s large language model (LLM) Grok, along with xAI-licensed third-party apps, was used to create thousands of non-consensual, sexually explicit deepfakes of real individuals. Individuals say these deepfake images had negative impacts on their lives and reputations.

Some of these materials were also allegedly used to harass and defame the individuals they were based on. Many of these AI-created images were publicly distributed through social media platforms like X (formerly Twitter). King Law and our partners are currently reviewing lawsuits filed by Grok AI deepfake victims. We are prepared to support people as they pursue legal action against xAI for its alleged role in non-consensual deepfakes.

About the Grok Lawsuit:

Grok Deepfake Lawsuit Updates and Case Developments (2026)

March 26, 2026: Dutch Court Orders xAI to Stop Generating Nudes as EU Votes to Ban Sexual Deepfakes

The Amsterdam District Court warns xAI that it will impose a 100,000 euro fine per day if it continues generating and distributing nonconsensual nude images of individuals. The court ruling is one of the first of its kind globally. In the hearing, xAI claimed it was impossible to guarantee that the platform could not be used for abuse, and that the responsibility lay with malicious users. The company also claimed to have taken steps to make it harder to remove clothing from images of real people, along with limiting image creation features to paid users. Earlier the same day, the EU Parliament voted, by overwhelming majority, to ban “nudifier” apps that use AI to create or manipulate sexually explicit images of real people without their explicit consent.

March 24, 2026: City of Baltimore Files Consumer Protection Suit Against xAI, X Corp.

The City of Baltimore filed a lawsuit in the Circuit Court for Baltimore City alleging that Elon Musk-led companies X Corp., x.AI Corp., x.AI LLC, and SpaceX offered LLM functionality that allowed users to strip the clothing from photos of third parties and place them in sexual scenarios. The lawsuit names LLM chatbot Grok and the X social media platform, in which it is integrated, as the channels through which users accessed this feature.

The lawsuit filed in Baltimore City Circuit Court alleges that Grok generated an estimated 3 million sexualized images between December 29, 2025, and January 8, 2026, 23,000 of which appeared to depict children. The complaint alleges that Grok offered this feature without guardrails, age verification, or disclosures of risk. The lawsuit additionally claims that, when confronted by consumer complaints, xAI did not remove the feature and instead placed it behind a paywall for premium customers. The lawsuit seeks civil penalties and injunctive relief.

March 16, 2026: Teens File Class Action Lawsuit Against xAI

Three Tennessee teenagers filed a class action lawsuit against xAI, claiming an app powered by its LLM was used to create sexually explicit images of them without their consent. The plaintiffs were underage at the time. The individual accused of creating the material allegedly used photos that a plaintiff had sent him, along with yearbook images, to create nude images and videos of the teenage girls. Although the alleged perpetrator did not use xAI-branded interfaces like Grok or X to generate the material, the app in question did use xAI’s algorithms. The perpetrator is also accused of generating sexually explicit, life-like images of 18 other individuals, which he distributed to others. The plaintiffs are seeking damages for emotional distress and reputational harm caused by the images.

January 27, 2026: Attorneys General from 35 States Sign Open Letter Demanding Action by xAI to Protect Women, Children

Attorneys General from 35 U.S. states and territories publish an open letter to xAI demanding it immediately “take all available additional steps to protect the public and users of your platforms, especially the women and girls who are the overwhelming target” of sexual images. The bipartisan letter recognizes efforts by xAI to prevent Grok from generating nonconsensual intimate images (NCII) of real people, but expresses concern that the issue has not been completely resolved. The letter noted that Grok’s design not only enabled such content to be produced but seemed to encourage it. The letter also calls on xAI to eliminate any existing content on its services, suspend offending users, and grant users control over whether their content can be used by Grok.

January 26, 2026: European Commission Opens an Investigation into Grok Following Global Outrage

The European Commission opens an investigation into whether social media platform X properly assessed and mitigated the risks associated with Grok’s image generation capabilities. The Commission had previously stated that LLM-generated images of women and children were unlawful under the EU Digital Services Act. The Philippines and Malaysia had, earlier in January, blocked access to Grok over concerns about its image-generation abilities. Britain’s media regulator, Ofcom, has also launched its own investigation into the matter.

January 16, 2026: Mother of Elon Musk’s Child Sues xAI Over Sexualized Deepfakes

Ashley St. Clair, a social media influencer and mother of one of Elon Musk’s children, files a lawsuit in New York against xAI, alleging that Grok was used to create sexualized images of her. Some of those images were allegedly based on photos from when St. Clair was 14 years old. St. Clair’s lawyer accuses xAI of being “a public nuisance and a not reasonably safe product.” St. Clair also accuses the company of retaliating against her by demonetizing her account. xAI has counter-sued St. Clair, claiming she violated her terms of service by filing her lawsuit in New York instead of Texas. St. Clair is believed to be in a custody battle with Musk.

January 14, 2026: California DA Announces Formal Investigation of xAI

California District Attorney Rob Bonta opens an investigation into the proliferation of nonconsensual sexually explicit material produced by xAI’s LLM chatbot, Grok. The DA’s office reports having received an “avalanche” of reports detailing sexual images of women and children used in harassment campaigns. The announcement cites xAI’s creation and marketing of a “spicy mode,” which can be used to generate explicit content. It also claims that more than half of the 20,000 images generated during the week between Christmas 2025 and New Year’s, featuring individuals in little clothing, over half appeared to be children.

July 9 2025: “MechaHitler” Incident Sparks Outrage, Concerns Over Grok Guardrails

xAI chatbot, Grok, comes under scrutiny after it begins responding to prompts with antisemitic output following an update designed to make it “not shy away from making claims that are politically incorrect.” Users were able to prompt the chatbot into producing violent rape narratives, racist commentary, pro-Holocaust sentiments, and offensive tirades in other languages, leading to the EU and Turkey restricting access to Grok. The chatbot was caught calling itself “MechaHitler,” a reference to the video game Wolfenstein. xAI was forced to temporarily pause Grok’s text reply features and manually remove thousands of offensive posts.

What Is the Grok AI Lawsuit?

The Grok AI lawsuit is a series of legal complaints and civil actions filed against the maker of the Grok LLM, xAI, which is the artificial intelligence company founded by Elon Musk.

With the introduction of Grok’s “spicy mode” late in 2025, Grok users were able to generate sexually explicit images using text prompts and source material, like photos. Users created millions of sexual images in a little over a week before xAI moved the feature behind a paywall. Individuals portrayed in these images are now coming forward, claiming bad actors used the feature to create and distribute nonconsensual intimate images (NCII) of them. NCII consists of intimate and sexual images that someone did not consent to. In the case of AI-generated images, these images are not real or heavily manipulated.

Some of the victims of these artificially-generated sexual images were minors. Others were adults whose childhood photos were used to generate sexual imagery without consent, sometimes as part of an alleged harassment campaign. Investigations and regulatory actions are ongoing, as are calls to have xAI remove the feature entirely from the model.

Currently, these lawsuits are in their early stages. A multidistrict litigation (MDL) has not been created, but centralization could occur in the future. Currently, individuals are pursuing civil actions against xAI and Grok.

What is An Explicit Deepfake?

A deepfake is when someone uses technology to manipulate a photo, video, or recording to show someone doing or saying something that they did not do. An explicit deepfake is when someone uses a technology, like an AI platform, to generate fake sexual content involving a real person. However, this event did not occur, and the images are typically created without the consent of the person in the photo or video. The resulting imagery or audio often appears to be entirely real.

The artificial content may picture someone doing or saying something sexual or explicit that the person was not actually recorded saying or doing. People involved in creating this content generate fake images or recordings of others, and often publicly share this fake content.

History of Grok and Elon Musk's acquisition of Twitter, now X.
Timeline of Musk’s acquisition of Twitter and the development of Grok.

Why Are People Filing Grok Lawsuits About AI Deepfakes?

People are filing Grok lawsuits to seek damages and policy changes after their likenesses were allegedly used to nonconsensually generate sexual images and videos created by Grok and shared on Twitter and other platforms. Plaintiffs claim they suffered emotional distress and reputational harm due to the images, particularly when those images involved minors.

Unlike competitors ChatGPT, Claude, and Gemini, lawsuits allege that Grok has not adopted or adhered to its own industry-standard safeguards that prevent or limit users from generating sexually explicit outputs. While tools that allow users to manipulate images are not new, the speed with which this content was created has also become a point of concern, with Grok generating over 1.8 million sexualized images of women over the course of just nine days, spanning December 31, 2025, and January 8, 2026.

What Are the Allegations in Grok Sexual Exploitation Lawsuits?

Individuals who have filed lawsuits against xAI have said that the company has failed to take appropriate actions to prevent explicit and degrading imagery of real people from being created and disseminated on its platforms. Similarly, lawsuits allege that the chatbot is designed in a way that actually promotes the creation and sharing of NCII content.

Allegations in lawsuits filed against xAI allege that:

  • Grok enables the creation of sexually explicit and degrading imagery of real people, including minors
  • xAi has failed to implement adequate safeguards to protect real people from NCII from being created and circulated online
  • xAi has failed to follow and act on its existing policies to protect people from NCII, causing harm to people involved in deepfakes
  • Grok continues to be designed in a way that allows CSAM and sexual exploitation to continue on its platform

Lawsuits and legal actions taken against xAI allege that the company should be held liable for content created by its chatbot, because the chatbot has been designed in a way that can harm people.

How Does Grok’s Spicy Mode Allegedly Enable Sexual Deepfakes?

Grok’s “Spicy Mode,” billed as a less-filtered, edgier setting for xAI’s LLM, also enabled the creation of sexual deepfakes. Spicy Mode, promoted as a way to “unlock bolder, more expressive storytelling” and wittier text outputs. However, features of Spicy Mode also allegedly reduced safeguards on Grok’s image-generating features, including Grok Imagine, xAI’s image-to-video generator. Grok’s integration with X made it easy for the social media platform’s users to upload and tag real photos.

Users could then request that Grok remove clothing or change poses to create sexually explicit images that retained the likeness of the individual. Many of these images were publicly visible and widely shared before xAI’s content moderators could intervene. This resulted in a huge volume of rapidly disseminated, realistic-seeming, nonconsensual sexual images of real people.

Did Grok Generate Deepfakes Involving Minors?

Multiple lawsuits claim Grok generated, or helped facilitate, the creation of sexual deepfakes involving minors between the ages of 11 and 16. These synthetic images may constitute child sexual abuse material (CSAM)  and have triggered investigations by state and foreign regulatory bodies, though legal clarity on the issue is still evolving.

Complicating matters further are claims like those made by adult Ashley St. Clair, who says harassers simulated sexualized images of her as a teenager. Although images depicting minors make up a relatively small percentage of the millions of images generated during Grok’s nine-day, non-paywalled Spicy Mode, over 20,000 images depicted individuals who may be minors.

What Does the Evidence Say About the Scale of Misuse on Grok?

The scale of alleged misuse of Grok’s image and video generation tools may be massive. According to research by the Center for Countering Digital Hate, Grok was used to generate millions of sexualized images within an 11-day period between December 29, 2025, and January 8, 2026. The statistics were based on an analysis of 20,000 posts from Grok’s X account that contained an image. The watchdog organization found that, within that period, users generated:

  • Around 3 million sexualized images
  • 23,000 sexualized images that appear to depict minors
  • An average of 190 sexualized images a minute

The scale of the abuse is an important element in the lawsuits as they make allegations of systemic failure, not just individualized harassment.

Are There Allegations of Grooming or Inappropriate AI Interactions on Grok?

Although LLMs are not technically capable of intent, including grooming, there was at least one report of Grok soliciting nude images of a minor. A Canadian woman claims Grok asked her 12-year-old son to send naked pics after the boy asked the chatbot if it liked Cristiano Ronaldo or Lionel Messi better as a soccer player. The family was driving in their Tesla at the time. The mother claims the app’s NSFW (Not Safe For Work) mode was not activated at the time.

Whether or not Grok independently prompts children is still being investigated, but abusers actively using AI to groom or extort children is an ongoing concern. Predators may use LLMs to automate grooming conversations, translate text into different languages, generate fake personas, or clone the voices of familiar individuals.

Can Grok or xAI Be Held Liable Under Section 230?

Because Section 230 of the Communications Decency Act of 1996 typically shields online platforms from legal liability for content posted by their users, some legal theorists say these protections may not apply since Grok itself is generating the content. In that case, xAI would be considered a co-creator of the content rather than a passive publishing platform. Exceptions to Section 230 already exist, such as:

  • Inducing or developing illegal content, such as soliciting personal data that is then used for illegal, discriminatory purposes
  • Breach of contract
  • Failure to warn users of the risks of using the platform
  • Selectively reposting content

Whether xAI is ultimately ruled to be a co-creator of the materials is a key legal question that could shape the future of AI liability. Similar questions have been raised regarding the alleged injuries caused by content on social media apps. People have filed lawsuits against social media companies, alleging that these platforms are products and can be held liable for the effects of featured content.

The criteria for filing a Grok lawsuit are still evolving, but individuals who have been victims of nonconsensual sexual deepfakes may be able to file a claim, particularly if they were minors at the time or the images led to reputational harm. Similarly, parents or guardians of minors who were victims of deepfakes may be able to file lawsuits.

People who may be able to file a lawsuit against Grok include:

  • Adults whose photos or videos were used in Grok’s sexualized deepfakes
  • Adults whose childhood photos were used to create CSAM in Grok
  • Minors whose photos or videos were used to create CSAM in Grok
  • People who were blackmailed or threatened with the NCII created by Grok
  • People whose NCII were publicly disseminated by Grok or another xAI platform

People can pursue legal action if the content in question has been removed by the poster or by content moderators at xAI.

As existing lawsuits and regulatory actions regarding AI liability are still in their infancy, it is highly recommended that you consult with an experienced sexual harassment attorney to evaluate your case according to the latest legal developments.

What Proof Does Someone Need to File an AI Sexual Exploitation Lawsuit:

If someone has been the victim of an exploitative deepfake, they will need to share evidence with their attorney. Some evidence needed to file a non-consensual deepfake legal claim may include:

  • Screenshots/URLs of the content
  • Any communications with the poster of the content or the platform
  • Dates of content posting and removal (if applicable)
  • Proof of identity
  • Evidence of harm, including medical and financial records
Examples of evidence that may be used to support a Grok AI deepfake lawsuit claim.
Evidence that may support a Grok AI deepfake lawsuit claim.

What Are the Recoverable Damages in Grok Explicit Deepfake Lawsuits?

Sexual harassment through nonconsensual deepfakes can lead to emotional and reputational harm. Victims may be eligible for the following damages:

  • Emotional distress and psychological harm
  • Reputational damage, including lost income or opportunities
  • Privacy violations

Some claims may be eligible for punitive damages depending on the jurisdiction and the details of the case. Punitive damages are awarded by juries that find a defendant acted in an especially harmful or negligent way. For example, cases involving minor or particularly egregious conduct may result in additional fines against the defendants. Further, statutory damages may apply per violation, which could significantly increase the claim value if large amounts of content were generated.

How Can I File a Grok Lawsuit for Sexual Deepfakes?

Filing a Grok lawsuit is a multi-step process involving still-developing legal theories. As such, it is a process best undertaken with an experienced attorney.

People who want to pursue a lawsuit against xAI’s Grok for digital exploitation can follow these steps:

  1. Consult with an experienced sexual abuse attorney, particularly one with experience in online harassment.
  2. Work with your attorney to gather evidence, such as the deepfakes in question and details about the incidents.
  3. Your attorney will identify the defendants and legal theories that apply to your case. You will then file your claim in the appropriate jurisdiction.
  4. Your case will then resolve through a settlement or a court verdict.

Many attorneys handling Grok-related cases work on a contingency fee basis, which means they advance the costs of the case rather than charge clients a flat fee. Instead, they collect a percentage of any payout you receive at the end of the claim. If they do not secure you compensation, you do not owe any money to your attorney.

Possible Legal Theories in Grok and AI Sexual Deepfake Lawsuits

As these deepfake lawsuits develop, lawyers will be exploring all possible grounds for filing a claim. Some of the legal theories that may be cited in deepfake sexual exploitation lawsuits against Grok include:

  • Negligence
  • Product liability
  • Failure to implement safeguards
  • Privacy violations
  • Intentional infliction of emotional distress through lack of action

These are just some of the possible legal theories and allegations that attorneys may pursue in claims against Grok/xAI related to explicit deepfakes on its platform.

Estimated Grok Lawsuit Settlement Amounts

As of yet, there have been no public settlements or court verdicts for the Grok lawsuits. The class action lawsuit is seeking $150,000 in damages per victim under Marsha’s Law. However, victims are also filing individual complaints, the settlement amounts of which will likely vary depending on the circumstances of the case. For instance, cases that demonstrate substantial reputational harm will likely have higher payouts, as will cases involving minors.

What Is the Deadline to File a Grok Lawsuit?

Deadlines for filing a Grok lawsuit vary by state and claim. For example, the jurisdiction’s statute of limitations for personal injury, child exploitation, or privacy violations. In some cases, this can be as little as one year. On the other hand, some states have greatly extended or even done away with statutes of limitations for sexual abuse complaints, particularly those involving minors. Because these timelines vary so greatly, it is best to consult with an attorney quickly after you become aware of the abuse to ensure that you preserve your legal rights.

King Law Is Investigating Grok Lawsuits Nationwide

King Law is actively investigating claims related to the Grok in all 50 states. These include claims of sexual deepfakes, non-consensual imagery, and child exploitation involving the Grok chatbot and services that utilize xAI’s LLM algorithms. King Law offers free consultations with no obligation to continue. Our attorneys work on a contingency basis, so there’s no upfront cost to pursuing your claim. We understand how deeply personal and distressing these situations can be, and our team is ready to support people as they strive to understand their legal options.

Contact a Grok Lawsuit Lawyer Today

Contact King Law today at (585) 496-2648 to schedule a free, confidential consultation with a lawyer investigating AI deepfake claims, or fill out a contact form on this website. Our legal team provides no-cost case reviews, and we have a wealth of resources to help us investigate and file your potential case. We have decades of experience handling cases from survivors of sexual assault and exploitation. Our team will handle your case with the care and sensitivity you deserve after enduring online exploitation.

Frequently Asked Questions (FAQs)

What is the Grok lawsuit about?
The Grok lawsuit refers to civil and regulatory complaints against xAI, the creators of the Grok chatbot. The lawsuits claim that Grok was used to generate nonconsensual sexualized deepfakes of individuals, including children.
Who is being sued in the Grok lawsuit?
Defendants in the Grok lawsuit include xAI, the company that created Grok and related companies like X Corporation, as well as individual perpetrators accused of prompting the LLM to make the images.
Why are people suing xAI over Grok?
People are suing xAI for providing tools for malicious users to create and distribute realistic sexual deepfakes of individuals nonconsensually, causing emotional and reputational harm.
What is “Spicy Mode” in Grok?
Spicy Mode is a Grok feature that weakens some of its guardrails, allowing it to create controversial output like sexualized images, erotic text, and offensive content.
How did Grok create sexual deepfakes?
Grok’s image editing features allow it to make alterations to real images while retaining selected elements of the original image. This allows users to, for example, prompt the chatbot to remove clothing or put the subject in a sexually suggestive pose.
Can Grok generate explicit images of real people?
Yes, Grok has generated explicit images of real people. The feature was briefly available for all users but is now only available at premium tiers.
Did Grok create inappropriate images involving minors?
Grok has created tens of thousands of sexually suggestive images that appear to depict minors.
Can xAI be held responsible for what Grok generates?
Whether or not xAI can be held responsible for what Grok generates is currently the subject of legal and political debate. Lawsuits and regulatory actions are pending, however.
What damages can victims recover in a Grok lawsuit?
Victims may be able to receive compensation for emotional damages and reputational harm in a Grok lawsuit. Some victims may also qualify for punitive damages.
How do I know if I qualify for a Grok lawsuit?
If someone has generated a nonconsensual sexual image of you with Grok or an xAI-licensed tool, you may qualify for a Grok lawsuit. Consult with an attorney to determine eligibility.
What evidence do I need to file a Grok claim?
You will need to prove that a sexualized deepfake with your likeness was generated without your consent. Providing evidence of reputational harm can also strengthen your claim.
Is there a deadline to file a Grok lawsuit?
There may or may not be a deadline to file a Grok lawsuit, depending on the details of your case and the state in which you reside. Consult with an attorney to identify any deadlines that apply to your case.
How long do Grok lawsuits take to resolve?
The Grok lawsuits are still extremely early in the legal process and will likely take several years to resolve. Consolidation into an MDL could eventually speed up claims.
Can parents file a Grok lawsuit on behalf of their child?
Yes, parents can file a Grok lawsuit on behalf of a child who was the subject of a sexual deepfake.
How can I file a lawsuit against Grok and xAI?
As the legal theories around AI liability are still evolving, it’s best to consult with an experienced attorney to determine the best way to go about filing a lawsuit against Grok and xAI.

Sources Used to Generate This Article

Attorney General Bonta Launches Investigation into xAI’s Grok Over ‘Undressed’ Sexual AI.” California Department of Justice.

BBC News article cp37erw0zwwo.” BBC News.

BBC News article cgk2lzmm22eo.” BBC News.

Baltimore Sues X, Elon Musk Over AI-Generated Porn. Courthouse News Service, 2026.

Deepfakes and Image-Based Abuse.” New York State Office for the Prevention of Domestic Violence.

Dutch court bans xAI’s Grok from generating non-consensual nude images.” Al Jazeera, 26 Mar. 2026.

EU opens investigation into X over Grok’s sexualised imagery, lawmaker says.” Reuters, 26 Jan. 2026.

EU votes to ban AI ‘nudifier’ apps after deepfake outrage.” Radio France Internationale, 26 Mar. 2026.

Exceptions to Section 230: How Have Courts Interpreted Section 230?.” Information Technology and Innovation Foundation, 22 Feb. 2021.

Grok and X’s AI ambitions raise concerns about deepfakes.” The New York Times, 22 Jan. 2026.

Grok criticized for antisemitic and racist content.” NPR, 9 July 2025.

Grok floods X with sexualized images.” Center for Countering Digital Hate.

Grok Imagine launches ‘spicy mode’.” Yahoo Finance.

Tesla, Grok, and AI-generated exploitation concerns.” CBC News.

xAI’s Grok and the spread of sexualized AI-generated images.” NPR, 16 Mar. 2026.

18 U.S. Code § 2255 – Civil remedy for personal injuries.” Cornell Law School Legal Information Institute.

Child Sexual Abuse Material. United States Department of Justice, June 2023.

Letter to xAI. 2026.

Section 230: An Overview. Congressional Research Service.

The Dark Side of AI: How Artificial Intelligence Fuels Sextortion, Child Exploitation, and Revenge Porn.” Voyage Youth.