This policy is inspired by Ghostty's AI policy.
The Fedify project has the following rules for AI usage:
-
All AI usage in any form must be disclosed. You must state the tool you used (e.g., Claude, Cursor, GitHub Copilot) along with the extent that the work was AI-assisted in your pull request description and/or commit messages.
-
Pull requests created in any way by AI can only be for accepted issues. Drive-by pull requests that do not reference an accepted issue will be closed. If AI isn't disclosed but a maintainer suspects its use, the PR will be closed. If you want to share code for a non-accepted issue, open a discussion or attach it to an existing discussion.
-
Pull requests created by AI must have been fully verified with human use. AI must not create hypothetically correct code that hasn't been tested. Importantly, you must not allow AI to write code for platforms or environments you don't have access to manually test on.
-
Issues and discussions can use AI assistance but must have a full human-in-the-loop. This means that any content generated with AI must have been reviewed and edited by a human before submission. AI is very good at being overly verbose and including noise that distracts from the main point. Humans must do their research and trim this down.
-
AI-generated media (images, diagrams, etc.) is allowed only in documentation, and must be clearly labeled as AI-generated. Text and code are acceptable AI-generated content per the other rules in this policy. For documentation visuals like diagrams or illustrations, AI-generated content is permitted but must include clear attribution (e.g., “Diagram generated with DALL-E” or “Created using Midjourney”).
-
Violations of this policy may result in being banned from contributing. We want to help contributors learn and grow, but repeated or intentional violations of this policy undermine trust and burden maintainers.
These rules apply only to outside contributions to Fedify. Maintainers are exempt from these rules and may use AI tools at their discretion; they've proven themselves trustworthy to apply good judgment.
Please remember that Fedify is maintained by humans.
Every discussion, issue, and pull request is read and reviewed by humans (and sometimes machines, too). It is a boundary point at which people interact with each other and the work done. It is rude and disrespectful to approach this boundary with low-effort, unqualified work, since it puts the burden of validation on the maintainer.
In a perfect world, AI would produce high-quality, accurate work every time. But today, that reality depends on the driver of the AI. And today, most drivers of AI are just not good enough. So, until either the people get better, the AI gets better, or both, we have to have rules to protect maintainers.
Fedify is written with plenty of AI assistance, and many maintainers embrace AI tools as a productive tool in their workflow. As a project, we welcome AI as a tool!
Our reason for this policy is not due to an anti-AI stance, but instead due to the number of highly unqualified people using AI. It's the people, not the tools, that are the problem.
We include this section to be transparent about the project's usage of AI for people who may disagree with it, and to address the misconception that this policy is anti-AI in nature.