X says it’s going to take action against creators who post AI videos of armed conflict without disclosure that the content is AI-generated. On Tuesday, X’s head of product, Nikita Bier, announced that people who use AI technology to mislead others in this way will be booted from the company’s Creator Revenue Sharing Program for a three-month period (90 days).
If they continue to post misleading AI content after the suspension lifts, they’ll be permanently suspended from the program.
“During times of war, it is critical that people have access to authentic information on the ground. With today’s AI technologies, it is trivial to create content that can mislead people,” Bier wrote on X. “Starting now, users who post AI-generated videos of an armed conflict — without adding a disclosure that it was made with AI — will be suspended from Creator Revenue Sharing for 90 days.”
X says it will identify the misleading posts through a combination of tools that are used to detect generative AI content, as well as through its crowdsourced fact-checking system, Community Notes.
X’s Creator Revenue Sharing Program offers creators the ability to generate income by posting on the platform and sharing in advertising revenue if their posts are popular. While designed to boost the amount of engaging content found on X, critics of the program say it incentivizes creators to post sensationalized content, like clickbait or other posts designed to spark outrage. Some have also criticized its lax content controls and its requirement that creators be paid X subscribers to participate.
Given how easy it is for AI to be used to make misleading photos and videos, X’s ban on financially rewarding creators for type of content is only a limited fix. Outside of war, AI media is often used to create political misinformation or push deceptive products in the influencer economy — all of which will still be allowed under the new policy.

