Skip to content

Consider changing stance on LLMs (detailed sources and explanations inside about gen AI) #36856

@ell1e

Description

@ell1e

Feature Description

I would suggest you consider your stance on generative AI/LLM code contributions:

All LLMs seem to inherently plagiarize heavily, even if you don't try to trick them into doing so: Source 1: https://dl.acm.org/doi/10.1145/3543507.3583199 Source 2: https://lcamtuf.substack.com/p/large-language-models-and-plagiarism Source 3 (this example seems to have been entirely without baiting the AI to do that! I find it shocking): https://www.pcgamer.com/software/ai/microsoft-uses-plagiarized-ai-slop-flowchart-to-explain-how-github-works-removes-it-after-original-creator-calls-it-out-careless-blatantly-amateuristic-and-lacking-any-ambition-to-put-it-gently/ Source 4: https://www.theatlantic.com/technology/2026/01/ai-memorization-research/685552/

This seems to be underestimated by a lot of gen AI users who seem to think that LLM plagiarism is a rare accident, rather than apparently the norm and apparently even when not baited to do so may concern larger parts of a single work.

And the training data seems to be broad and questionable: https://archive.is/1EzVK https://www.anthropic.com/transparency https://www.theatlantic.com/technology/archive/2025/03/libgen-meta-openai/682093/

(I'm not a lawyer, this isn't legal advice. Best check the many article sources I'm giving here yourself and form your own opinion.)

AI seems to be potentially damaging to open-source as a whole as well:

Vibe Coding Is Killing Open Source Software, Researchers Argue

The study Vibe Coding Kills Open Source [...] asks the question: is vibe coding economically sustainable? Can OSS survive when so many of its users are takers and not givers? According to the study, no.

“[...] under traditional OSS business models, where maintainers primarily monetize direct user engagement… higher adoption of vibe coding reduces OSS [...] welfare,” the study said.

This is already happening. [...] Tailwind Labs is extremely popular, more popular than it’s ever been, but revenue has plunged.

Overrun with AI slop, cURL scraps bug bounties to ensure “intact mental health”

The project developer for one of the Internet’s most popular networking tools is scrapping its vulnerability reward program after being overrun by a spike in the submission of low-quality reports, much of it AI-generated slop. [...] AI slop has already flooded music-streaming services with so many songs—often misattributed to real artists—that the platforms are slowly becoming unusable for music discovery. cURL’s move may be an early indication that something similar is happening to bug bounty programs.

It also seems to be potentially damaging to engineers who use it:

Relying on AI in colonoscopies risks eroding doctors’ skills, study warns

Routine use of AI technology can cause the ability of specialists to detect precancerous growths to decline by a fifth, authors conclude.

ChatGPT May Be Eroding Critical Thinking Skills, According to a New MIT Study

Researchers used an EEG to record the writers’ brain activity across 32 regions, and found that of the three groups, ChatGPT users had the lowest brain engagement and “consistently underperformed at neural, linguistic, and behavioral levels.”

Does using artificial intelligence assistance accelerate skill decay and hinder skill development without performers’ awareness?

The available evidence suggests that frequent engagement with automation induces skill decay. Given that (a) AI is often designed to take over more advanced cognitive processes than non-AI automation, and (b) skill decay is accelerated for cognitive skills, AI-induced skill decay is a likely consequence of frequent engagement with AI assistants.

And the Apple study has shown apparently LLMs lack logical thought, which seems both weird regarding plagiarism, since how can you be transformative without being able to do any coherent thought, and regarding software quality:

Intelligencec Illusion: What Apple’s AI Study Reveals About Reasoning

The Apple research team's findings are both methodical and damning. [...] both model types experienced complete collapse when faced with high-complexity tasks. [...] Rather than improving with increased complexity [...] these models showed a peculiar pattern: their reasoning effort would increase up to a certain point, then decline dramatically despite having adequate computational resources. This suggests that the models weren’t actually reasoning at all— they were following learned patterns that broke down when confronted with novel challenges. [...] When the veneer of sophisticated language is stripped away, what remains is a sophisticated but ultimately hollow mimicry of thought.

Apparently AIs tend to make coders think they're faster too while that apparently not actually being true:

What Actually Happens When Programmers Use AI Is Hilarious, According to a New Study

As flagged by Ars Technica, a new study from the nonprofit Model Evaluation and Threat Research (METR) found that in practice, programmers are actually slower when using AI assistance tools than making do without them.

[...] the METR team found that when using AI tools, their subjects did indeed spend less time actively coding, debugging, researching, or testing — but that was because they instead spent their time “reviewing AI outputs, prompting AI systems, and waiting for AI generations.”

Anyway, I'm not an expert, read the sources for yourself and don't trust my conclusions.

But based on this, might be worth to question LLM use and what licensing LLM code output can even reasonably have, and whether it should be banned from gitea.

Even if a ban cannot be easily enforced in some situations, I assume e.g. manual copy&pasting code without the proper licensing is also banned, so the enforcement shouldn't necessarily prevent a ban.

Other projects that already banned gen AI: Gentoo Servo Loupe Qemu postmarketOS GoToSocial NetBSD Zig

Screenshots

No response

Metadata

Metadata

Assignees

No one assigned

    Labels

    type/proposalThe new feature has not been accepted yet but needs to be discussed first.

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions