General
George Miloradovich
Researcher, Copywriter & Usecase Interviewer
February 25, 2025
A low-code platform blending no-code simplicity with full-code power 🚀
Get started free
February 25, 2025
•
7
min read

Content Boundaries: Can Grok-2 Generate NSFW Images and How It's Regulated

George Miloradovich
Researcher, Copywriter & Usecase Interviewer
Table of contents

Grok-2 is an AI image generator with fewer restrictions on NSFW (Not Safe For Work) content compared to competitors like DALL-E and Midjourney. This leniency has sparked concerns about potential misuse, legal risks, and ethical issues. Here's a quick breakdown:

  • Grok-2: Minimal NSFW filtering, no watermarks, limited copyright protection, and no deepfake prevention.
  • DALL-E: Strict NSFW filters, watermarked images, advanced prompt screening, and robust safeguards.
  • Midjourney: PG-13 policy, automated moderation, community oversight, and user reporting tools.

Quick Comparison

Platform NSFW Filtering Watermarking Copyright Protection Deepfake Prevention
Grok-2 Minimal None Limited None
DALL-E Very Strict Yes Strong Yes
Midjourney Strict (PG-13 Policy) Yes Moderate Yes

While Grok-2 allows more creative freedom, its lack of safeguards raises serious concerns. For safer and more controlled outputs, DALL-E and Midjourney are better choices.

Midjourney vs DALL·E 3: Ultimate Comparison

Midjourney

1. Grok-2 Safety Standards

Grok-2

Grok-2 takes a different route when it comes to content moderation. Although it claims to avoid pornographic, overly violent, or deceptive images, as well as copyright violations, these rules seem loosely applied. Several operational shortcomings stand out:

  • Content Filtering: Compared to the strict NSFW filters used by major platforms, Grok-2 applies far fewer restrictions.
  • Image Watermarking: AI-generated images lack watermarks, leaving users without clear indicators of their origin.
  • Copyright Protection: The platform offers limited safeguards, falling short of the robust copyright verification processes common in the industry.
  • Deepfake Prevention: There are no measures in place to prevent deepfakes, which raises concerns about the potential for manipulated imagery.

These gaps have already drawn attention from legal and policy experts.

"Grok and Gemini are two AI image generators that generate controversial images but for the wrong reasons. Both generators can be used to create false and misleading images, Gemini because of too much filtering and Grok because of too little filtering"

"This is one of the most reckless and irresponsible AI implementations I've ever seen"

Despite the presence of automated classifiers and safety tools, Grok-2's measures fall short of industry expectations. This raises serious legal and ethical questions about the platform's approach to managing content.

2. DALL-E Safety Standards

DALL-E

OpenAI's DALL-E uses a layered safety system to minimize the creation of inappropriate images. This approach relies on advanced filtering techniques to block problematic content effectively.

The platform's safety measures include:

  • Content Classification: Sophisticated classifiers guide the model away from harmful content .
  • Training Data Filtering: Explicit material is removed from the training data .
  • Prompt Screening: Prompts that mention public figures are automatically rejected .
  • Image Watermarking: All generated images are marked with AI-origin watermarks .

However, these safeguards aren't foolproof. Researchers have demonstrated that methods like the 'SneakyPrompt' jailbreak can bypass these defenses .

"Our work basically shows that these existing guardrails are insufficient. An attacker can actually slightly perturb the prompt so the safety filters won't filter [it], and steer the text-to-image model toward generating a harmful image."
– Neil Zhenqiang Gong, Assistant Professor at Duke University

OpenAI has responded quickly to such challenges, updating the system to block bypass attempts .

"Safety is our priority and we take a multi-pronged approach. In the underlying DALL-E 3 model, we've worked to filter the most explicit content from its training data including graphic sexual and violent content, and have developed robust image classifiers that steer the model away from generating harmful images. We've also implemented additional safeguards for our products, ChatGPT and the DALL-E API – including declining requests that ask for a public figure by name. We identify and refuse messages that violate our policies and filter all generated images before they are shown to the user. We use external expert red teaming to test for misuse and strengthen our safeguards."

While some users have criticized these restrictions for limiting artistic freedom , OpenAI's focus remains on responsible AI use. These safety measures demonstrate a clear commitment to ethical deployment, which will be further contrasted in the upcoming platform comparison.

sbb-itb-23997f1

3. Midjourney Safety Standards

Midjourney has put in place strict content moderation rules to ensure its platform remains Safe For Work (SFW). This is achieved through a mix of automated tools and community-led oversight.

The platform uses several safety measures, including advanced algorithms to detect problematic prompts , automatic blocking of inappropriate text and image inputs, community moderators who review flagged content, and a user reporting system to address violations .

Midjourney follows a PG-13 policy, which bans the creation of NSFW content such as nudity, sexual imagery, and fetish material. Breaking these rules can lead to temporary or permanent account bans . The platform explains:

"Midjourney is an open-by-default community. To keep the platform accessible and welcoming to the broadest number of users, content must be 'Safe For Work' (SFW)."

However, maintaining this balance comes with challenges. Some users and artists have raised concerns about the platform's restrictions. AI artist and author Tim Boucher highlights a key issue:

"If the user is a consenting adult, the problem is reduced. On the other hand, as a Trust & Safety professional, your system should not be creating nudes when people aren't asking for it... Especially since your rules officially disallow nudes. And when users ask for it directly, they may be banned outright from using the service. There's a major disconnect here."

Midjourney faces hurdles such as overly strict filters that sometimes block harmless content , accidental creation of inappropriate material despite safeguards , and users finding ways to bypass restrictions with alternative keywords .

To address these issues, the platform regularly updates its systems to block workarounds and relies on active community moderation. This positions Midjourney as a middle ground between more lenient platforms like Grok-2 and heavily restrictive alternatives .

Platform Safety Comparison

Safety features differ significantly across leading AI image generation platforms, particularly in how effectively they prevent misleading or harmful content. NewsGuard's testing highlights these differences :

Platform False Content Generation Rate Content Restrictions Watermarking Primary Safety Features
Grok-2 80% Minimal None Lacks clear safety measures
Midjourney 45% Strict Yes AI-driven moderation, community oversight
DALL-E 3 10% Very strict Yes Advanced prompt filtering, built-in safeguards

Technical Implementation Differences

The way each platform handles safety is shaped by its technical design. DALL-E 3 stands out as the most restrictive, leveraging ChatGPT and GPT-4 to better interpret and block problematic prompts . For example, when users request content involving public figures or sensitive topics, the system often generates neutral alternatives instead.

Rob Rosenberg, a legal consultant at Telluride Legal Strategies, highlights a key issue with Grok-2:

"Unlike other platforms that label their AI-generated images with a watermark that identifies them as such, Grok does not tag its image results in any way that would clue in downstream customers as to their origin. This puts such downstream customers at risk of being misled – misinterpreting such images as genuine and not a deepfake."

Safety Implementation Comparison

The safety measures reflect each platform's priorities and philosophy:

  • DALL-E 3: Adopts strict policies, rejecting prompts about public figures and sensitive topics to maintain a high level of control .
  • Midjourney: Relies on AI-based moderation and community oversight to analyze and filter content .
  • Grok-2: Operates with minimal restrictions, allowing content that other platforms would block.

These differences highlight the ongoing balance between promoting creative freedom and ensuring responsible AI use. While Grok-2 prioritizes fewer restrictions in the name of free expression, this approach raises concerns about the potential misuse of its tools and the legal risks tied to controversial image generation.

Summary and Recommendations

Our analysis of platform moderation highlights clear differences that affect both usability and legal exposure. For users prioritizing strict oversight, Midjourney offers a more controlled environment with detailed safety measures. On the other hand, Grok-2 provides more freedom but requires users to take additional precautions.

Use Case Recommended Platform Key Considerations
Responsible Image Generation Midjourney Strong moderation tools with context-aware safeguards
Unrestricted Generation Grok-2 Limited moderation; allows for broader creative possibilities

These findings emphasize the importance of organizational strategies for managing AI-generated content. Based on these comparisons, organizations should consider the following steps:

  • Policy Updates: Revise AI usage policies to address risks associated with unmoderated outputs .
  • Content Moderation: Implement strict filtering systems to prevent harmful or inappropriate content .
  • Watermarking: Use clear watermarks to mark AI-generated images .

Matt Hasan, founder and CEO of aiRESULTS, stresses the importance of oversight:

"The absence of robust oversight mechanisms means that harmful or unethical content could proliferate unchecked. To mitigate these issues, implementing stringent content moderation policies, developing advanced detection tools to identify manipulated images and establishing clear legal frameworks for AI-generated content are crucial steps. Also, fostering transparency in AI development and encouraging industry-wide standards can help ensure that such technologies are used responsibly and ethically."

For independent creators and small teams, Grok-2's flexibility can be appealing, but it requires heightened awareness and careful monitoring .

Related Blog Posts

Application

Try now

Related Blogs

Use case

Backed by