u/Aware_Activity_6591

Normally, AI systems detect moderation violations and display an error message. Seedance 2.0 appears to handle this differently. When I tested boundary cases, content was generated—but not as expected. Instead of a direct refusal, the video showed a character explaining that the request couldn't be fulfilled. When I intentionally pushed further, the behavior changed. Sometimes only a "moderated" message appeared. However, upon repeating the same prompt, a video was generated featuring American police officers in the background, with the character explaining that the request violates the terms of service.

Is this behavior intentional, or a bug? To me, it seems like a potentially clever feature—a way to make refusals more "explainable" and harder to manipulate through prompts. If intentional, this could be an interesting approach to protecting AI systems against jailbreaking attempts.

reddit.com
u/Aware_Activity_6591 — 15 days ago