General

George Miloradovich
Researcher, Copywriter & Usecase Interviewer
February 25, 2025
Grok-2 is an AI image generator with fewer restrictions on NSFW (Not Safe For Work) content compared to competitors like DALL-E and Midjourney. This leniency has sparked concerns about potential misuse, legal risks, and ethical issues. Here's a quick breakdown:
Platform | NSFW Filtering | Watermarking | Copyright Protection | Deepfake Prevention |
---|---|---|---|---|
Grok-2 | Minimal | None | Limited | None |
DALL-E | Very Strict | Yes | Strong | Yes |
Midjourney | Strict (PG-13 Policy) | Yes | Moderate | Yes |
While Grok-2 allows more creative freedom, its lack of safeguards raises serious concerns. For safer and more controlled outputs, DALL-E and Midjourney are better choices.
Grok-2 takes a different route when it comes to content moderation. Although it claims to avoid pornographic, overly violent, or deceptive images, as well as copyright violations, these rules seem loosely applied. Several operational shortcomings stand out:
These gaps have already drawn attention from legal and policy experts.
"Grok and Gemini are two AI image generators that generate controversial images but for the wrong reasons. Both generators can be used to create false and misleading images, Gemini because of too much filtering and Grok because of too little filtering"
"This is one of the most reckless and irresponsible AI implementations I've ever seen"
Despite the presence of automated classifiers and safety tools, Grok-2's measures fall short of industry expectations. This raises serious legal and ethical questions about the platform's approach to managing content.
OpenAI's DALL-E uses a layered safety system to minimize the creation of inappropriate images. This approach relies on advanced filtering techniques to block problematic content effectively.
The platform's safety measures include:
However, these safeguards aren't foolproof. Researchers have demonstrated that methods like the 'SneakyPrompt' jailbreak can bypass these defenses .
"Our work basically shows that these existing guardrails are insufficient. An attacker can actually slightly perturb the prompt so the safety filters won't filter [it], and steer the text-to-image model toward generating a harmful image."
– Neil Zhenqiang Gong, Assistant Professor at Duke University
OpenAI has responded quickly to such challenges, updating the system to block bypass attempts .
"Safety is our priority and we take a multi-pronged approach. In the underlying DALL-E 3 model, we've worked to filter the most explicit content from its training data including graphic sexual and violent content, and have developed robust image classifiers that steer the model away from generating harmful images. We've also implemented additional safeguards for our products, ChatGPT and the DALL-E API – including declining requests that ask for a public figure by name. We identify and refuse messages that violate our policies and filter all generated images before they are shown to the user. We use external expert red teaming to test for misuse and strengthen our safeguards."
While some users have criticized these restrictions for limiting artistic freedom , OpenAI's focus remains on responsible AI use. These safety measures demonstrate a clear commitment to ethical deployment, which will be further contrasted in the upcoming platform comparison.
Midjourney has put in place strict content moderation rules to ensure its platform remains Safe For Work (SFW). This is achieved through a mix of automated tools and community-led oversight.
The platform uses several safety measures, including advanced algorithms to detect problematic prompts , automatic blocking of inappropriate text and image inputs, community moderators who review flagged content, and a user reporting system to address violations .
Midjourney follows a PG-13 policy, which bans the creation of NSFW content such as nudity, sexual imagery, and fetish material. Breaking these rules can lead to temporary or permanent account bans . The platform explains:
"Midjourney is an open-by-default community. To keep the platform accessible and welcoming to the broadest number of users, content must be 'Safe For Work' (SFW)."
However, maintaining this balance comes with challenges. Some users and artists have raised concerns about the platform's restrictions. AI artist and author Tim Boucher highlights a key issue:
"If the user is a consenting adult, the problem is reduced. On the other hand, as a Trust & Safety professional, your system should not be creating nudes when people aren't asking for it... Especially since your rules officially disallow nudes. And when users ask for it directly, they may be banned outright from using the service. There's a major disconnect here."
Midjourney faces hurdles such as overly strict filters that sometimes block harmless content , accidental creation of inappropriate material despite safeguards , and users finding ways to bypass restrictions with alternative keywords .
To address these issues, the platform regularly updates its systems to block workarounds and relies on active community moderation. This positions Midjourney as a middle ground between more lenient platforms like Grok-2 and heavily restrictive alternatives .
Safety features differ significantly across leading AI image generation platforms, particularly in how effectively they prevent misleading or harmful content. NewsGuard's testing highlights these differences :
Platform | False Content Generation Rate | Content Restrictions | Watermarking | Primary Safety Features |
---|---|---|---|---|
Grok-2 | 80% | Minimal | None | Lacks clear safety measures |
Midjourney | 45% | Strict | Yes | AI-driven moderation, community oversight |
DALL-E 3 | 10% | Very strict | Yes | Advanced prompt filtering, built-in safeguards |
The way each platform handles safety is shaped by its technical design. DALL-E 3 stands out as the most restrictive, leveraging ChatGPT and GPT-4 to better interpret and block problematic prompts . For example, when users request content involving public figures or sensitive topics, the system often generates neutral alternatives instead.
Rob Rosenberg, a legal consultant at Telluride Legal Strategies, highlights a key issue with Grok-2:
"Unlike other platforms that label their AI-generated images with a watermark that identifies them as such, Grok does not tag its image results in any way that would clue in downstream customers as to their origin. This puts such downstream customers at risk of being misled – misinterpreting such images as genuine and not a deepfake."
The safety measures reflect each platform's priorities and philosophy:
These differences highlight the ongoing balance between promoting creative freedom and ensuring responsible AI use. While Grok-2 prioritizes fewer restrictions in the name of free expression, this approach raises concerns about the potential misuse of its tools and the legal risks tied to controversial image generation.
Our analysis of platform moderation highlights clear differences that affect both usability and legal exposure. For users prioritizing strict oversight, Midjourney offers a more controlled environment with detailed safety measures. On the other hand, Grok-2 provides more freedom but requires users to take additional precautions.
Use Case | Recommended Platform | Key Considerations |
---|---|---|
Responsible Image Generation | Midjourney | Strong moderation tools with context-aware safeguards |
Unrestricted Generation | Grok-2 | Limited moderation; allows for broader creative possibilities |
These findings emphasize the importance of organizational strategies for managing AI-generated content. Based on these comparisons, organizations should consider the following steps:
Matt Hasan, founder and CEO of aiRESULTS, stresses the importance of oversight:
"The absence of robust oversight mechanisms means that harmful or unethical content could proliferate unchecked. To mitigate these issues, implementing stringent content moderation policies, developing advanced detection tools to identify manipulated images and establishing clear legal frameworks for AI-generated content are crucial steps. Also, fostering transparency in AI development and encouraging industry-wide standards can help ensure that such technologies are used responsibly and ethically."
For independent creators and small teams, Grok-2's flexibility can be appealing, but it requires heightened awareness and careful monitoring .