Our Expert in Greece
No results available
As the EU AI Act’s high-risk and transparency obligations take full effect on 2 August 2026, intellectual property lawyers Greece-wide are fielding an unprecedented volume of compliance questions from platforms, publishers, creative agencies and rights-holders. Article 50 of the EU AI Act introduces binding transparency rules that touch every organisation developing, deploying or distributing AI-generated content in or into the Greek market. At the same time, Greece’s existing copyright framework, anchored in Law 2121/1993, creates distinct liability exposures for companies that train models on protected works without authorisation. This guide maps the Article 50 obligations onto the Greek legal landscape, explains training-data copyright risks, and provides a practical enforcement and compliance playbook for businesses operating in Greece.
The EU Artificial Intelligence Act (Regulation (EU) 2024/1689) entered into force on 1 August 2024. Its phased implementation reaches a critical milestone on 2 August 2026, when transparency obligations under Article 50 and the full regime for high-risk AI systems become enforceable. Any organisation that provides, deploys or integrates an AI system that generates or manipulates text, images, audio or video, or that interacts directly with natural persons, must comply.
Greek businesses are directly affected because the AI Act is a regulation, not a directive: it applies without the need for transposition into national law. In-house counsel at technology companies, platform operators, publishers, advertising agencies, music labels and collecting societies should treat 2 August 2026 as a hard deadline. Rights-holders who suspect their works have been used to train generative AI models without a licence also gain new evidentiary tools under the transparency framework.
The three highest-priority actions for Greek organisations right now are: (1) audit all training datasets for copyrighted material and document licensing status; (2) implement machine-readable labelling and watermarking for every synthetic output; and (3) update supplier contracts, terms of service and internal incident-response procedures to reflect Article 50 requirements.
Quick Checklist, 5 First Actions Before 2 August 2026
Article 50 of the EU AI Act establishes transparency obligations that apply regardless of whether an AI system is classified as high-risk. Its core purpose is to ensure that individuals know when they are interacting with an AI system or consuming AI-generated content, and that synthetic outputs can be detected by downstream actors and rights-holders.
Under Article 50, providers of AI systems designed to interact directly with natural persons must ensure that individuals are informed they are interacting with an AI system, unless this is obvious from the circumstances. Providers of AI systems that generate synthetic audio, image, video or text content must ensure that the outputs are marked in a machine-readable format and are detectable as artificially generated or manipulated. Deployers who use AI systems to generate or manipulate content that constitutes a deepfake must disclose that the content has been artificially generated or manipulated. The European Commission’s Code of Practice on marking and labelling of AI-generated content provides supplementary technical guidance on how to implement these requirements in practice.
| Entity | Key Article 50 Obligations (Transparency) | Practical Examples / Immediate Action |
|---|---|---|
| Providers (developers) | Ensure AI outputs are marked in a machine-readable format; provide technical documentation; embed metadata that enables detection of synthetic content | Integrate watermarking into model pipelines; maintain auditable logs of training datasets; publish model cards with transparency disclosures |
| Deployers (integrators / platforms) | Inform users when they interact with an AI system; label AI-generated content presented to end-users; conduct risk assessments for high-risk use cases | Add clear UI notices (“This content was generated by AI”); update terms of service; implement content-moderation rules for synthetic outputs |
| Online platforms / marketplaces | Implement marking and labelling policies for AI-generated content hosted on the platform; cooperate with rights-holders on takedown requests | Deploy platform-level labelling badges; create dedicated rights-holder reporting channels; retain metadata for evidentiary use |
The transparency requirements under the EU AI Act are not merely aspirational. Non-compliance exposes providers and deployers to administrative fines and, in Greece, to supplementary civil liability where a failure to label or disclose contributes to copyright infringement or consumer harm.
Article 50 is a regulatory transparency obligation, it does not, by itself, create or extinguish copyright. However, it materially changes the enforcement landscape for rights-holders in Greece by generating new sources of evidence and new grounds for claiming that an infringer acted knowingly.
Greece’s primary copyright statute is Law 2121/1993 on Copyright, Related Rights and Cultural Matters. The law protects original works of authorship and requires that a work be the “personal intellectual creation” of its author. This human-authorship requirement means that purely AI-generated content, produced without meaningful human creative input, is unlikely to qualify for copyright protection under Greek law. Where a human author uses AI as a tool but exercises genuine creative choices in selecting, arranging or editing the output, the resulting work may qualify for protection, though the boundaries remain untested in Greek case law.
Greece transposed Directive (EU) 2019/790 (the DSM Directive) into national law, including provisions on text and data mining (TDM). The TDM exception permits reproductions and extractions for the purposes of text and data mining by research organisations and cultural heritage institutions (Article 3 of the Directive), and a broader TDM exception exists under Article 4, but only where rights-holders have not expressly reserved their rights in a machine-readable manner. This opt-out mechanism is critical: rights-holders in Greece who fail to implement machine-readable reservations on their content may inadvertently allow their works to be mined for AI training purposes.
Article 50’s transparency framework strengthens the position of Greek rights-holders in several ways. First, the obligation on providers to maintain training-data documentation gives rights-holders a concrete evidentiary basis for demanding disclosure of whether their works were used in training. Second, machine-readable labelling of synthetic outputs makes it easier to detect when AI-generated material competes with or copies protected works. Third, the combination of Article 50 disclosures and Greece’s existing copyright law framework creates a pathway for injunctive relief where providers fail to demonstrate lawful use of training data. Industry observers expect that Greek courts will increasingly treat a failure to comply with Article 50 transparency duties as relevant to assessments of fault and damages in copyright liability actions.
The question of whether companies may train AI on copyrighted works without permission is central to intellectual property lawyers Greece practitioners are advising on today. The short answer under Greek law is: no blanket permission exists, and unauthorised use of protected works for model training carries real infringement risk.
Under Law 2121/1993, reproduction and adaptation of a copyrighted work require the consent of the rights-holder unless a specific statutory exception applies. The TDM exceptions transposed from the DSM Directive provide limited safe harbours, but Article 4’s broader TDM exception is subject to a rights-holder opt-out. Where a rights-holder has placed a machine-readable reservation (such as a robots.txt directive, a metadata tag or a contractual restriction), the exception does not apply, and any reproduction for AI training purposes constitutes infringement unless independently licensed. Academic research on TDM and AI in the European context confirms that the opt-out mechanism is the decisive variable in determining whether training-data use is lawful.
For Greek businesses developing or fine-tuning AI models, this means that every dataset component must be traced to a clear legal basis: either a licence, a statutory exception that has not been opted out of, or public-domain status.
A rigorous training-data audit should document the following fields for every data source:
When licensing training data from third-party suppliers, Greek businesses should insist on contractual provisions that allocate risk clearly. Sample clauses to consider include:
On the technical side, watermarking AI content and embedding provenance metadata at the point of generation assist with both Article 50 compliance and downstream copyright enforcement. The trade-off is implementation cost and processing overhead, but the legal risk of non-compliance increasingly outweighs these costs.
Determining copyright liability in Greece for infringement caused by generative AI involves multiple potential defendants and overlapping legal bases. The AI Act’s transparency regime does not create a new cause of action for copyright infringement, but it materially affects how claims are investigated, pleaded and proved.
Under Law 2121/1993, any person who reproduces, adapts, distributes or communicates a protected work to the public without authorisation may be liable. In the generative-AI context, this could include:
Greek courts apply the general tort provisions of the Civil Code (Articles 914 et seq.) alongside the specific remedies in Law 2121/1993. Remedies available to rights-holders include injunctive relief (including interim measures), compensatory damages, disgorgement of profits, destruction of infringing copies, and publication of the judgment.
IP enforcement in Greece follows several channels, each with distinct procedural characteristics:
Rights-holders who suspect that their works have been used to train an AI model or that AI-generated outputs infringe their copyright should take the following steps:
Because AI models and their outputs routinely cross borders, Greek rights-holders frequently need to coordinate enforcement across multiple EU jurisdictions. The Brussels I bis Regulation (Regulation (EU) 1215/2012) governs jurisdiction in cross-border intellectual property disputes, and Greek courts may assert jurisdiction where the infringing content is directed at or accessible in Greece. The AI Act’s uniform transparency obligations across all Member States simplify the evidentiary burden: a provider’s failure to comply with Article 50 in any EU jurisdiction is equally actionable in Greece.
Achieving compliance with Article 50 and avoiding copyright liability Greece-based organisations face requires a structured, step-by-step approach. The following playbook is designed for in-house legal and compliance teams preparing for the 2 August 2026 enforcement date.
Step 1, Scope and classify your AI systems. Inventory every AI system in use across the organisation. Classify each system according to the AI Act’s risk categories: prohibited practices, high-risk, limited-risk (transparency obligations under Article 50), and minimal-risk. Any system that generates synthetic text, images, audio or video falls at minimum into the Article 50 transparency category.
Step 2, Conduct a training-data audit. For each AI system, map every dataset used for training, fine-tuning and validation. Record the fields described in the audit checklist above. Flag any dataset where the rights-holder’s opt-out status is unclear or where no licence is on file.
Step 3, Update contracts and licences. Revise supplier agreements to include training-data licence grants, rights-clearance warranties and indemnities. Update customer-facing terms of service to disclose the use of AI-generated content and the nature of any synthetic outputs.
Step 4, Implement labelling, watermarking and machine-readable metadata. Deploy technical measures that mark every AI-generated output as synthetic. Options include invisible digital watermarks, C2PA (Coalition for Content Provenance and Authenticity) metadata and IPTC photo metadata fields. The European Commission’s Code of Practice on marking and labelling of AI-generated content provides sector-specific implementation guidance.
Step 5, Update user notices and interface disclosures. Ensure that every user-facing interaction with an AI system includes a clear, timely and accessible disclosure. For chatbots, this means a notice at the start of the conversation. For content platforms, this means a visible label on every AI-generated item.
Step 6, Establish rights-holder reporting and evidence-preservation processes. Create a dedicated channel for rights-holders to report suspected training-data infringement or output-level copying. Define internal SLAs for acknowledging, investigating and resolving complaints. Retain all relevant logs and metadata as evidence.
Step 7, Build an incident-response and remediation plan. Prepare a playbook for responding to enforcement actions, regulatory inquiries and rights-holder litigation. Designate a cross-functional response team (legal, engineering, communications) and conduct tabletop exercises before the August 2026 deadline.
| Milestone | Date | Action Required |
|---|---|---|
| AI Act entered into force | 1 August 2024 | Begin legal analysis and system inventory |
| Prohibited-practices ban enforceable | 2 February 2025 | Cease any prohibited AI practices |
| Transparency & high-risk obligations enforceable | 2 August 2026 | Full Article 50 compliance; labelling, documentation and reporting channels operational |
Selecting the right technical approach depends on the type of content generated and the organisation’s existing technology stack:
When selecting a vendor or building in-house capability, prioritise interoperability with the C2PA standard, robustness against adversarial removal, and compatibility with the European Commission’s code of practice.
The intersection of the EU AI Act and Greek copyright law demands specialist legal guidance. Intellectual property practitioners in Greece offer services spanning the full compliance and enforcement lifecycle: AI-system classification and risk assessment; training-data audits and licensing strategy; drafting and negotiating data-supply and model-deployment agreements; implementing Article 50 transparency procedures; representing rights-holders in infringement actions (including emergency injunctive relief); and defending providers and deployers against copyright and regulatory claims.
Given the complexity and the immovable 2 August 2026 deadline, early engagement with experienced intellectual property lawyers Greece businesses can rely on is not optional, it is a risk-management imperative. Whether you are a start-up fine-tuning a foundation model, a publisher protecting a catalogue, or a platform integrating third-party AI, specialist counsel can identify exposures before they become liabilities.
This article was produced by Global Law Experts. For specialist advice on this topic, contact Maria Athanassiadou at Dr. Helen G. Papaconstantinou and Partners Law Firm, a member of the Global Law Experts network.
posted 4 minutes ago
posted 27 minutes ago
posted 28 minutes ago
posted 51 minutes ago
posted 57 minutes ago
posted 1 hour ago
posted 2 hours ago
posted 2 hours ago
posted 2 hours ago
posted 2 hours ago
posted 3 hours ago
posted 3 hours ago
No results available
Find the right Advisory Expert for your business
Sign up for the latest advisor briefings and news within Global Advisory Experts’ community, as well as a whole host of features, editorial and conference updates direct to your email inbox.
Naturally you can unsubscribe at any time.
Global Law Experts is dedicated to providing exceptional legal services to clients around the world. With a vast network of highly skilled and experienced lawyers, we are committed to delivering innovative and tailored solutions to meet the diverse needs of our clients in various jurisdictions.
Global Advisory Experts is dedicated to providing exceptional advisory services to clients around the world. With a vast network of highly skilled and experienced advisors, we are committed to delivering innovative and tailored solutions to meet the diverse needs of our clients in various jurisdictions.
Send welcome message