No-you-still-dont-have-to-pay-but-any-AI-use-is-restricted-license. 1 This way, everyone knows, its not as free as the MIT license and has absolutely no relation to it. /s
Ofcourse you can specialize existing licenses with limited paragraphs and reflect that in their names...
Key words are:
- permission is [...] granted
- free of charge
- without restriction
- use, copy, …
Then:> may not be used for the purposes of […]
The license contradicts itself.
> Don't we have to ask for permission before feeding someone's years of work into an AI?
That's the point of an OpenSource license, to give permission.
This kind of stuff makes me think very few people really understand what OpenSource is about. The very same people who will fallback to licenses such as the BSL as soon as people/companies will use the permissions that they gave, and then will complain that "no one wants to pay for the thing i did for free and nobody asked for".
Providers of GPAI models must respect Text and Data Mining (TDM) opt-outs.
2.1 Legal Basis: Article 53(1)(c) AI Act and Directive (EU) 2019/790 The Copyright Chapter of the Code directly addresses one of the most contentious legal questions in AI governance: the use of copyrighted material in training GPAI models and the risk of infringing outputs. Article 53(1)(c) AI Act requires GPAI providers to “identify and respect copyright protection and rights reservations” within their datasets. + This obligation complements the framework of Directive (EU) 2019/790 on copyright and related rights in the Digital Single Market (DSM Directive). Notably, Article 4(3) DSM Directive allows rightsholders to exclude their works from text and data mining (TDM) operations via machine-readable opt-outs.
https://www.ddg.fr/actualite/the-european-unions-code-of-pra...
https://www.europarl.europa.eu/RegData/etudes/STUD/2025/7740...
Two relevant bits, dug out from the 175-page whole:
> Although the Act tries to address this by extending obligations to any provider placing a GPAI model on the EU market, the extraterritorial enforcement of these obligations remains highly uncertain due to the territorial nature of copyright law and the practical difficulty of pursuing infringement claims when training occurs under foreign legal standards, such as U.S. fair use.
and:
> Finally, it is important to clarify that the current EU framework provides a closed list of exceptions and does not recognise a general fair use defence. As a result, AI-generated outputs that include protected expression without a valid exception remain unlawful.
It seems to be dated the same month as DDG's analysis, July 2025, so I would expect the MIT Non-AI License that we're discussing here to be much more defensible in the EU than in the U.S. — as long as one studies that full 175-page "Generative AI and Copyright" analysis and ensures that it addresses the salient points necessary to apply and enforce in EU copyright terms. (Queued for my someday-future :)
It’s midnight now, so you’re on your own to dig up and review specific instances of relevant case law, or to contrast with non-U.S. laws. Licensing above refers to i.e. LICENSE files of the specific sort that this post is about ("MIT Non-AI License"); other definitions of licensing, as well as e.g. DMCA exceptions, exist that might be of interest for you to explore further. I believe there’s been a handful of cases related to AI and fair use this past year, but as with all such defenses, unique circumstances are common enough that I hesitate to suggest any future outcome as 100% certain without much more case law than AI has today. (I am not your lawyer, this is not legal advice.)
But can I give you another "viewpoint"? I guess it's like, "Wow, my code, my work, what came from my brain, my fingers" - it essentially lives forever, if you think about it - it becomes embedded and compressed inside weights/tokens. Like - part of you is there.
I guess it's cool. For me it's just to know that like this super intelligent things deep down actually knows who I am - my code is in it's architecture, it gives me feeling of honor in some way. You know?
Just my take.
You deserve to be recognized beyond the false religion of the singularity.
Literature
Without references and citations—- never mind.
Go read Animal Farm
- VC funded SF startups with an ethos of "Do whatever is necessary at first, scale then lawyer up once a unicorn and shits starts to hit the fan" this might try to prevent... and who will blissfully ignore it
- actually mindful actors, e.g public researchers, non-profit, etc who genuinely understand the point and will respect it but whom, ironically enough, you might want to support by providing them with more resources including your code
So... yes the intent is good but honestly I wouldn't trust the 1st category of actors to respect that. Technically what they "should" do, because it's safe and efficient, is only crawl the most popular compatible licenses, e.g. MIT, and ignore the rest. That's safe and pragmatic. Again what I expect them to do (just my hunch) is take EVERYTHING, get called on, apologize, change nothing, get brought to court, apologize again and only do something if the financial repercussion is bigger than the expected alternative. So probably still do nothing.
We should move to local LLMs.
Related, would you be in violation if you hosted this in a public GitHub repo, since it's in the TOS that they use source for training AI?
https://www.ddg.fr/actualite/the-european-unions-code-of-pra...
If you want to release your code as actual open source but legally restrict what AI companies do with it, use a copyleft license like the GPL. They could still use it for training, but the product of such training may itself fall under the GPL being a derivative work, and the AI corps don't want to touch that legal quagmire. (The GPL continues to be Stallman's brilliant legal aikido at work. Stallman is like that one guy from Final Fantasy Tactics Advance who gives you the power to manipulate the laws to work to your advantage.)
Honestly, it may be time to abandon open source as a concept. There are source-available strategies that cause less social harm than open source, including Coraline Ada Ehmke's "ethical source" and the Sentry project's "fair source" (https://fair.io) models that we can draw inspiration from.