Site iconSite icon ForkLog

AI-generated content is not yours

AI-generated content is not yours

As generative artificial-intelligence tools radically reshape how content is created and consumed, questions of copyright, intellectual property and data ownership are moving to the fore.

In a ForkLog column, Tatiana Kontareva, a lawyer at Aurum, analyses the key regulatory aspects of using and distributing AI-created content and outlines strategies to help Web3 projects ensure legal safety and build transparent relationships with users.

Who owns AI-generated content?

Despite rapid progress in AI, the question remains open and hotly debated in legal and technology circles. The creation process typically involves several actors: model developers, trainers, users and the AI models themselves. Yet when it comes to ownership of the final output, there is still no single, definitive answer as to who—if anyone—has a lawful claim.

If you build or use generative-AI tools, this bears directly on your business model, monetisation and exposure to legal risks. When technology outpaces rule-making, legal strategy is not merely important; it can prove decisive for sustainable growth and the project’s legal safety.

In essence, projects integrating LLMs and other AI tools should keep two points in mind:

From our observations and experience, most companies integrating LLMs and other AI tools—including industry leaders such as OpenAI and Google (Gemini)—generally do not claim ownership of outputs generated by their models. However, while title to final materials may not be a priority, projects typically seek to retain the right to use user prompts and AI-generated materials to further train models. At this stage, intellectual property becomes especially significant—not only legally, but also for user trust and the ethical handling of data in the development and deployment of AI technologies.

Consider two main categories of data most often used to train models:

Training data: what can and cannot be used

Most copyright disputes arise from using protected materials to train AI models without proper permission. To work effectively and realise their full potential—especially LLMs—models need large and diverse datasets. This creates a tension between the need for breadth and the constraints of copyright law.

First and foremost, it is crucial to know where your model obtains its training data. As a rule, AI-based platforms draw on at least two main sources:

Does “fair use” shield against copyright infringement?

In some instances, the law does allow the use of copyright-protected content without a licence—but only if certain conditions are met. One of the best known is the doctrine of “fair use” (fair use). In The New York Times v OpenAI, the latter argued that training models on publicly available content falls under this doctrine. It is not absolute, however, and its application requires careful legal assessment case by case.

Courts typically weigh four factors when deciding whether use is fair:

In sum, the risk of infringement rises markedly if potentially infringing content is repeatedly used in training or reproduced in generated outputs—especially where there are no effective mechanisms to monitor, identify and remove such content, even after a possible violation is flagged. Projects must therefore ensure that their training datasets and practices comply with applicable copyright law.

Practical legal strategies for deploying AI tools

Launching AI products, or products in which AI is a critical component, is not only a matter of technical innovation; it requires thorough legal and operational planning. Below are key points to reduce risk and make the product legally sound:

Conclusions and key takeaways

As AI technologies advance at unprecedented speed, the legal, ethical and regulatory questions surrounding the training, commercialisation and deployment of AI systems grow more complex. Key takeaways for projects focused on integrating and using models in their products:

For founders, developers and business leaders working with AI and Web3 technologies, complying with applicable law is not just a box to tick; it is integral to a successful strategy. A competent legal approach can protect the business, build user trust, enhance model reliability and support the long-term viability of innovation.

Exit mobile version