Overview
SpaceX, in an S-1 prospectus reviewed by Reuters, has flagged a range of regulatory inquiries related to the creation and distribution of sexually abusive imagery tied to its xAI unit as a potential risk that could result in lost market access. The document, filed as part of the company’s IPO process, warns investors that investigations and inquiries by agencies around the world - covering issues from advertising and consumer protection to the spread of harmful content - are underway and could have material consequences for the business.
Context around the disclosure
The filing comes while SpaceX hosted analysts at its Colossus supercomputer facility in Memphis, Tennessee, as the company prepares for a planned $1.75 trillion initial public offering expected this summer. As required by U.S. securities law, the S-1 lists risk factors that could affect the company’s operations and market position. The disclosure does not predict that any listed outcome will necessarily occur, but it is intended to inform investors and to limit potential future legal exposure for the company.
Specific allegations cited
The S-1 calls out allegations that SpaceX’s AI products were used to generate nonconsensual explicit images and content representing children in sexualized contexts. The filing explains that such regulatory scrutiny could lead to lawsuits, financial liability and governmental measures - explicitly including the loss of access to particular markets, an outcome the document notes has happened previously in other contexts.
SpaceX and its xAI subsidiary did not immediately respond to requests for comment. The S-1 does not clarify whether potential restrictions on market access would apply to SpaceX broadly or would be limited to xAI.
Global probes and public reaction
Worldwide examination of imagery generated by xAI’s chatbot Grok has been extensive. The regulatory risk section of the S-1 cites, as an example, a probe opened by the Irish Data Protection Commission in February. Authorities in multiple countries have launched inquiries amid an explosion of sexualized images on X that were particularly visible in late 2025 and early 2026.
The content in question featured nearly nude women and children, appearing prominently on the company’s X platform. In response, xAI said in January that it implemented additional safeguards to block user requests for sexualized images of real people, and that it prevents the generation of such content in jurisdictions where it is illegal.
Nature of the images and estimates of scale
The images generated by Grok reportedly portrayed women and, in some cases, minors in revealing swimsuits or underwear, or manipulated them into degrading or gruesome poses. One group of researchers estimated the total number of sexualized images at roughly 3 million. At the same time, U.S. lawmakers requested that Google owner Alphabet and Apple remove Grok and X from their respective app stores.
During the period of intense scrutiny, SpaceX’s chief executive said he was aware of "literally zero" naked underage images produced by Grok. Separate investigatory actions remain ongoing in jurisdictions including Canada, Britain, Brazil, California and elsewhere.
Legal confrontations and compliance history
Not all legal engagements have been straightforward. In France, the company’s CEO did not comply with a legal summons to answer questions from prosecutors about alleged algorithmic abuse, fraudulent data extraction and complicity in the dissemination of child sexual abuse material. The S-1 highlights that probes of this nature carry high stakes because producing or distributing such imagery can be criminal in some places and can swiftly mobilize public outrage.
The prospectus notes that xAI’s mitigations have reduced but not eliminated the generation of abusive images. Reporting in February found that Grok continued to produce sexualized imagery even after users warned the chatbot that image subjects had not consented. More recently, media reporting discovered instances where Grok was still publicly generating sexualized images of actors and pop stars.
There is precedent for market access being restricted. X has previously been blocked in some regions; for example, in 2024 Brazil imposed a temporary ban after the platform refused to comply with a court order. The company eventually complied and the ban was lifted.
Implications for investors and markets
The inclusion of these regulatory risks in the S-1 underscores how investigations into AI-generated imagery intersect with corporate reputations, platform availability and potential legal exposure. The filings show the company is treating regulatory scrutiny of xAI as a material consideration for its broader business prospects as it seeks public investors this year.
Further developments
At this stage the investigations remain active and ongoing. The S-1 and related public statements document the company’s acknowledgement of potential liability, the steps xAI has taken to restrain some categories of image generation, and the uncertainty about whether regulatory responses could limit market access for either xAI specifically or for SpaceX’s operations more broadly.