Normally, AI systems detect moderation violations and display an error message. Seedance 2.0 appears to handle this differently. When I tested boundary cases, content was generated—but not as expected. Instead of a direct refusal, the video showed a character explaining that the request couldn't be fulfilled. When I intentionally pushed further, the behavior changed. Sometimes only a "moderated" message appeared. However, upon repeating the same prompt, a video was generated featuring American police officers in the background, with the character explaining that the request violates the terms of service.
Is this behavior intentional, or a bug? To me, it seems like a potentially clever feature—a way to make refusals more "explainable" and harder to manipulate through prompts. If intentional, this could be an interesting approach to protecting AI systems against jailbreaking attempts.