AIs.txt
AIs.txt
is a mental model of a machine learning permission system.
Intuitively, question this: what if we could make a human-readable file that declines machine learning (a.k.a. Copilot use)?
It's like robots.txt, but for Copilot.
User-agent: OpenAI
Disallow: /some-proprietary-codebase/
User-agent: Facebook
Disallow: /no-way-mark/
User-agent: Copilot
Disallow: /expensive-code/
Sitemap: /public/sitemap.xml
Sourcemap: /src/source.js.map
License: MIT
# SOME LONG LEGAL STATEMENTS HERE
Key Issues
Would it be legally binding?
For now, no. It would be a polite way to mark my preference to opt out of such data mining. It's closer to the Ask BigTechs Not to Track option rather than a legal license. Technically, Apple's App Tracking Transparency does not ban all tracking activity; it never can.
Why not LICENSE
or COPYING.txt
?
Both are mainly written in human language and cannot provide granular scraping permissions depending on the collector.
Also, GitHub Copilot ignores LICENSE
or COPYING.txt
,
claiming we consented to Copilot using our codes for machine learning by signing up and pushing code to GitHub,
We may expand the LICENSE
system to include the terms for machine learning use,
but that would even more edge case and chaotic licensing systems.