I am concerned about LLM code https://github.com/python/cpython/commit/951675c18a1f97513f495b9ec604054e0702eaaf in Python. This isn’t legal advice. I’ll merely link and quote you the sources that made me concerned, and you can draw your own conclusions.
Here’s a video clip of apparently a lawyer, reviewing what looks like plagiarism of a single source, triggered by e.g. function isEven() {:
https://github.com/mastodon/mastodon/issues/38072#issuecomment-4105681567
Here’s a high profile incident https://www.pcgamer.com/software/ai/microsoft-uses-plagiarized-ai-slop-flowchart-to-explain-how-github-works-removes-it-after-original-creator-calls-it-out-careless-blatantly-amateuristic-and-lacking-any-ambition-to-put-it-gently/ concerning Microsoft themselves.
The following study seems to suggest memorization may correlate with model performance https://www.sciencedirect.com/science/article/pii/S2949719123000213#b7:
We found that the models that consistently output the highest-quality text are also the ones that have the highest memorization rate.
This seems to suggest, together with study on lack of reasoning ability https://machinelearning.apple.com/research/illusion-of-thinking, that LLMs can’t draw their own conclusions:
We found that LRMs have limitations in exact computation: they fail to use explicit algorithms and reason inconsistently across puzzles. We also investigate the reasoning traces in more depth, […] ultimately raising crucial questions about their true reasoning capabilities.
Takeaway of Forbes: “even the most sophisticated reasoning models fundamentally lack genuine cognitive abilities.”
Takeaway of The Atlantic: “Large language models don’t “learn”—they copy.”
There’s also this field study https://dl.acm.org/doi/10.1145/3543507.3583199 suggesting a plagiarism rate of gen AI of at least 2-5%, and that’s the part they could pin down.
It seems like above sources suggest LLMs plagiarizing the training data may be common, may happen even when not baited, and may concern single identifiable sources at significant length. But check the sources yourself.
I’m guessing the AI user will usually be unaware whenever this happens, therefore likely unable to stop it.
Therefore, I suggest even just for ethics and respecting the FOSS license of other projects, Python should consider banning LLM code submissions.