You just published a blog post written with ChatGPT. Or an Instagram reel with an AI voiceover. Or a product photo where AI swapped the background.
Do you have to label any of that?
The short answer: it depends. The long answer involves the EU AI Act, platform-specific rules, and national laws that are still catching up.
In this article, I break down all three levels (national, EU-wide, and technology companies) so you know exactly what applies to you, when, and how.
- The EU AI Act's labeling requirement for AI-generated content is in effect. Violations can cost up to €15 million or 3% of annual revenue.
- National laws don't exist yet in most countries, but platforms like YouTube and Meta have already introduced their own labeling requirements
- Labeling must be "clear and conspicuous". For deepfakes and publicly relevant topics, there is an absolute labeling requirement.
1. National Labeling Requirements
In most countries, there are currently no national guidelines or laws requiring the labeling of AI-generated content.
A labeling requirement has been discussed for some time, however. In Germany, for example, the Left Party submitted a motion in mid-June 2023 in the Bundestag demanding a labeling requirement when using AI systems to generate content "that is made publicly accessible or is not purely private in nature." However, this motion was rejected in November 2023.
The Federal Commissioner for Data Protection (BfDI) also emphasizes in a statement the importance of labeling AI content as a protective mechanism for consumers, particularly in the context of propaganda and defamatory content.
Additionally:
Existing regulations on consumer protection, personal rights, or competition law could potentially still apply.
2. EU AI Act

The EU launched the Artificial Intelligence Act in March 2024, a directive to legally establish EU-wide regulations for AI systems.
In Article 50, it also provides transparency obligations for providers and users of certain AI systems and tools:
- Providers of AI tools that generate audio, image, video, or text content must ensure that tool outputs are labeled in a machine-readable format as artificially generated or manipulated.
- Users of AI image generators, AI voice generators, AI music generators, or AI video generators must disclose when they publish deepfakes (deceptively realistic AI-generated image, audio, or video content of people, places, or events).
- Users must also label AI texts on topics of public interest as artificially generated. Exception: The text has been reviewed and editorially processed by humans.
In Recitals 132-136, the necessity of labeling AI-generated content is explained in more detail.
It emphasizes that users must be able to recognize AI-generated content, which is increasingly difficult to distinguish from human-made content. The labeling requirement is intended to counter the risk of misinformation and manipulation.
Recital 134 clarifies that the labeling requirement for deepfakes should not impair freedom of expression and artistic freedom. For clearly artistic, satirical, or fictional works, only appropriate disclosure of AI use is required without affecting the work itself.
The labeling requirement becomes binding once the AI regulation is fully transposed into national law. The EU AI Act has been in force since August 2024, and transposition into national law is expected to be completed by mid-2026:
2.1 When Does the EU AI Act Come Into Effect?
The timeline for the EU AI Act taking effect can be summarized as follows:
- On March 13, 2024, the European Parliament approved the AI Act, formally adopting it.
- On May 21, 2024, the EU Council approved the law. Details can be found in the press release.
- 20 days after publication in the Official Journal of the EU, the regulation enters into force.
- Most provisions of the AI Act will then apply 24 months after entry into force, so presumably from mid-2026.
However, some provisions have shorter transition periods:
- Bans on particularly risky AI systems apply after just 6 months
- Rules for generative AI apply after 12 months
- Obligations for "embedded AI systems" only after 36 months
This means the EU AI Act entered into force in August 2024, but most regulations do not apply until mid-2026. Some particularly important provisions, such as bans on high-risk systems, took effect much earlier.
2.2 How Must AI Content Be Labeled According to the EU AI Act?
The EU AI Act contains no specific requirements for exactly how AI-generated content must be labeled.
The AI Act merely prescribes that AI-generated content such as texts, images, audio, and video must be labeled "clearly" as such. Users should be able to clearly recognize when content has been machine-generated.
Particularly in marketing activities using AI content, it must be clearly disclosed that these are AI creations.
When publishing AI-generated texts intended to inform the public about important topics, it must also be made transparent that the texts were artificially created.
However, how this transparency and labeling is to be specifically implemented, e.g., through disclaimer texts, watermarks on images, notes in metadata, etc., is left open by the AI Act.
The exact design of the labeling requirement will likely be left again to providers and users of AI systems, meaning legal clarity will only be achieved through numerous court rulings (similar to GDPR).
2.3 What Are the Consequences for Violating the Labeling Requirement?
Companies that violate the labeling requirement for AI-generated content under the EU AI Act could face the following consequences:
1. High Fines
For violations of obligations under the AI Act, such as the labeling requirement, fines of up to €15 million or 3% of global annual turnover can be imposed, whichever is higher.
However, lower caps are expected to apply to small and medium-sized enterprises as well as startups.
2. Ban on AI System
Providers may be forced to remove their AI systems from the market if they repeatedly or seriously violate the regulations.
3. Competition Law Consequences
In addition to fines under the EU AI Act, competition law consequences may also be threatened, e.g., cease and desist letters.
3. Labeling on Major Online Platforms
Some major online platforms and social networks have already voluntarily and before the introduction of the EU AI Act introduced a labeling requirement for creators and/or automatically label AI content. Here's a brief overview:
3.1 YouTube
Since November 2023, creators must indicate when uploading YouTube videos whether they contain AI-generated content that could appear real.
The labeling is done via an option in YouTube Studio called "Altered content":

The labeling is displayed in the video description:

For sensitive topics like news or elections, YouTube displays the notice directly in the video:

Violations can according to YouTube Help lead to removal of videos, exclusion from the Partner Program, or result in a permanent label being displayed in videos that cannot be removed by creators.
3.2 Meta (Facebook, Instagram, Threads)
Meta has announced that from May 2024, it will label a broader range of AI-generated content such as videos, audios, and images with the "Made with AI" label:

These changes are based on recommendations from the Oversight Board as well as extensive consultations with experts and public opinion surveys.
The goal is to create more transparency and context for users, rather than simply removing such content and potentially restricting freedom of expression.
3.3 TikTok
TikTok introduced in September 2023 a new tool that allows creators to label their AI-generated content:

The goal is to create transparency and inform TikTok users when content has been created or edited with AI.
According to the press release, the company is also testing automatic labeling for AI-generated content and working with experts to develop effective labeling guidelines and implement industry best practices for transparency and responsible innovation in AI.
3.4 Twitter/X
X (formerly Twitter) currently has no specific recommendations, guidelines, or requirements for labeling AI-generated content.






