this post was submitted on 12 Apr 2026
658 points (95.3% liked)

Technology

84019 readers
3212 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
top 50 comments
sorted by: hot top controversial new old
[–] Blue_Morpho@lemmy.world 280 points 1 week ago (7 children)

The title of the article is extraordinary wrong that makes it click bait.

There is no "yes to copilot"

It is only a formalization of what Linux said before: All AI is fine but a human is ultimately responsible.

" AI agents cannot use the legally binding "Signed-off-by" tag, requiring instead a new "Assisted-by" tag for transparency"

The only mention of copilot was this:

"developers using Copilot or ChatGPT can't genuinely guarantee the provenance of what they are submitting"

This remains a problem that the new guidelines don't resolve. Because even using AI as a tool and having a human review it still means the code the LLM output could have come from non GPL sources.

[–] marlowe221@lemmy.world 76 points 1 week ago* (last edited 1 week ago) (1 children)

Yeah, that’s also my question. Partially because I am a former-lawyer-turned-software-developer… but, yeah. How are the kernel maintainers supposed to evaluate whether a particular PR contains non-GPL code?

Granted, this was potentially an issue before LLMs too, but nowhere near the scale it will be now.

(In the interests of full disclosure, my legal career had nothing to do with IP law or software licensing - I did public interest law).

[–] wonderingwanderer@sopuli.xyz 12 points 1 week ago (17 children)

If it's flagged as "assisted by " then it's easy to identify where that code came from. If a commercial LLM is trained on proprietary code, that's on the AI company, not on the developer who used the LLM to write code. Unless they can somehow prove that the developer had access to said proprietary code and was able to personally exploit it.

If AI companies are claiming "fair use," and it holds up in court, then there's no way in hell open-source developers should be held accountable when closed-source snippets magically appear in AI-assisted code.

Granted, I am not a lawyer, and this is not legal advice. I think it's better to avoid using AI-written code in general. At most use it to generate boilerplate, and maybe add a layer to security audits (not as a replacement for what's already being done).

But if an LLM regurgitates closed-source code from its training data, I just can't see any way how that would be the developer's fault...

load more comments (17 replies)
load more comments (6 replies)
[–] theherk@lemmy.world 138 points 1 week ago (11 children)

Seems like a reasonable approach. Make people be accountable for the code they submit, no matter the tools used.

[–] ell1e@leminal.space 27 points 1 week ago (2 children)

If the accountability cannot be practically fulfilled, the reasonable policy becomes a ban.

What good is it to say "oh yeah you can submit LLM code, if you agree to be sued for it later instead of us"? I'm not a lawyer and this isn't legal advice, but sometimes I feel like that's what the Linux Foundation policy says.

[–] ViatorOmnium@piefed.social 52 points 1 week ago (11 children)

But this was already the case. When someone submitted code to Linux they always had to assume responsibility for the legality of the submitted code, that's one of the points of mandatory Signed-off-by.

load more comments (11 replies)
load more comments (1 replies)
load more comments (10 replies)
[–] sonofearth@lemmy.world 63 points 1 week ago (6 children)

I am the c/fuck_ai person but at this point I have made peace we can't avoid it. I still don't want it to do artsy stuff (image gen, video gen) and to blindly use it in critical stuff because humans are the ones that should be doing it or have constant oversight. I think the team's logic is correct here, because there is no way to know if the code is from an LLM or a human unless something there screams LLM or the contributor explicitly mentions it. Mandating the latter seems like a reasonable move for now.

load more comments (6 replies)
[–] catlover@sh.itjust.works 58 points 1 week ago (5 children)

I'd still be highly sceptical about pull requests with code created by llms. Personally what I noticed is that the author of such pr doesn't even read the code, and i have to go through all the slop

[–] kcuf@lemmy.world 20 points 1 week ago (2 children)

Ya I'm finding myself being the bad code generator at work as I'm scattered across so many things at the moment due to attrition and AI can do a lot of the boilerplate work, but it's such a time and energy sink to fully review what it generates and I've found basic things I missed that others catch and shows the sloppiness. I usually take pride in my code, but I have no attachment to what's generated and that's exposing issues with trying to scale out using this

[–] Repelle@lemmy.world 18 points 1 week ago* (last edited 1 week ago) (2 children)

Same. There’s reduction in workforce, pressure to move faster, and no good way to do that without sloppiness. I have never been this down on the industry before; it was never great, but now it’s terrible.

load more comments (2 replies)
load more comments (1 replies)
load more comments (4 replies)
[–] CanIFishHere@lemmy.ca 52 points 1 week ago (3 children)

AI is here, another tool to use...the correct way. Very reasonable approach from Torvalds.

[–] Newsteinleo@infosec.pub 30 points 1 week ago (8 children)

I don't have a problem with LLMs as much as the way people use them. My boss has offloaded all of his thinking to LLMs to the point he can't fix a sentence in a slide deck without using an LLM.

It's the people that try to use LLMs for things outside their domain of expertise that really cause the problems.

load more comments (8 replies)
load more comments (2 replies)
[–] NewNewAugustEast@lemmy.zip 52 points 1 week ago* (last edited 1 week ago) (4 children)

Copilot? You mean the AI with terms of service that are in bold and explicit: "for entertainment purposes only"?

Which is why its in the title and not the article? EntertainBait?

[–] Zacryon@feddit.org 22 points 1 week ago (10 children)

I suppose GitHub Copilot is meant, which is a different thing.

load more comments (10 replies)
load more comments (3 replies)
[–] SethTaylor@lemmy.world 41 points 1 week ago* (last edited 1 week ago) (8 children)

Bad actors submitting garbage code aren't going to read the documentation anyway, so the kernel should focus on holding human developers accountable rather than trying to police the software they run on their local machines.

"Guns don't kill people. People kill people"

Torvalds and the maintainers are acknowledging reality: developers are going to use AI tools to code faster, and trying to ban them is like trying to ban a specific brand of keyboard.

The author should elaborate on how exactly AI is like "a specific brand of keyboard". Last I checked a keyboard only enters what I type, without hallucinating 50 extra pages. And if AI, a tool that generates content, is like "a specific brand of keyboard", does that mean my brain is also a "specific brand of keyboard"?

I get their point. If you want to create good code by having AI create bad code and then spending twice the time to fix it, feel free to do that. But I'm in favor of a complete ban.

[–] Simulation6@sopuli.xyz 44 points 1 week ago

The keyboard thing is sort of a parable, it is as difficult to determine if code was generated in part by AI as it is to determine what keyboard was used to create it.

[–] Miaou@jlai.lu 33 points 1 week ago

The (very obvious) point is that this cannot be enforced. So might as well deal with it upfront.

[–] Shayeta@feddit.org 18 points 1 week ago* (last edited 1 week ago) (2 children)

AI is a useful tool for coding as long as it's being used properly. The problem isn't the tool, the problem is the companies who scraped the entire internet, trained LLM models, and then put them behind paywalls with no options to download the weights so that they could be self-hosted. Brazen, unaccountable profiteering off of the goodwill of many open source projects without giving anything back.

If LLMs were community-trained on available, open-source code with weights freely available for anyone to host there wouldn't be nearly as much animosity against the tech itself. The enemy isn't the tool, but the ones who built the tool at the expense of everyone and are hogging all the benefits.

load more comments (2 replies)
load more comments (5 replies)
[–] webkitten@piefed.social 39 points 1 week ago (3 children)

Linux kernel being written by Microsoft's AI.

[–] MoogleMaestro@lemmy.zip 18 points 1 week ago

Microsoft needs to try to ruin Linux somehow, it can't just hurt windows 11 with AI slop code, it needs to expand it's efforts to other systems.

load more comments (2 replies)
[–] null@lemmy.org 38 points 1 week ago (4 children)

Ah, the solution that recognizes there's no way to eliminate AI from the supply chain after it's already been introduced.

load more comments (4 replies)
[–] 0ndead@infosec.pub 37 points 1 week ago (20 children)

“Yes to Copilot, no to AI slop”

Pick One

load more comments (20 replies)
[–] gandalf_der_12te@discuss.tchncs.de 33 points 1 week ago* (last edited 1 week ago) (1 children)

I agree. If AI becomes outlawed, it will simply be used without other people knowing about it.

This approach, at least, means that people will label AI-generated code as such.

[–] emmy67@lemmy.world 18 points 1 week ago

Maybe. There's still strong disapproval around it. I can imagine many will still hide it.

[–] Jankatarch@lemmy.world 28 points 1 week ago (1 children)

Maintainers' only responsibility is to ensure quality and shouldn't have to check for rogue AI submissions.

Tho I still miss consistent fucking weather so year of the netbsd?

load more comments (1 replies)
[–] peacefulpixel@lemmy.world 27 points 1 week ago

"yes to copilot no to AI slop" lol lmfao

[–] hperrin@lemmy.ca 26 points 1 week ago

There are so many reasons not to include any AI generated code.

https://sciactive.com/human-contribution-policy/#Reasoning

[–] ell1e@leminal.space 26 points 1 week ago* (last edited 1 week ago) (30 children)

Ultimately, the policy legally anchors every single line of AI-generated code

How would that even be possible? Given the state of things:

https://dl.acm.org/doi/10.1145/3543507.3583199

Our results suggest that [...] three types of plagiarism widely exist in LMs beyond memorization, [...] Given that a majority of LMs’ training data is scraped from the Web without informing content owners, their reiteration of words, phrases, and even core ideas from training sets into generated texts has ethical implications. Their patterns are likely to exacerbate as both the size of LMs and their training data increase, [...] Plagiarized content can also contain individuals’ personal and sensitive information.

https://www.theatlantic.com/technology/2026/01/ai-memorization-research/685552/

Four popular large language models—OpenAI’s GPT, Anthropic’s Claude, Google’s Gemini, and xAI’s Grok—have stored large portions of some of the books they’ve been trained on, and can reproduce long excerpts from those books. [...] This phenomenon has been called “memorization,” and AI companies have long denied that it happens on a large scale. [...]The Stanford study proves that there are such copies in AI models, and it is just the latest of several studies to do so.

https://www.twobirds.com/en/insights/2025/landmark-ruling-of-the-munich-regional-court-(gema-v-openai)-on-copyright-and-ai-training

The court confirmed that training large language models will generally fall within the scope of application of the text and data mining barriers, [...] the court found that the reproduction of the disputed song lyrics in the models does not constitute text and data mining, as text and data mining aims at the evaluation of information such as abstract syntactic regulations, common terms and semantic relationships, whereas the memorisation of the song lyrics at issue exceeds such an evaluation and is therefore not mere text and data mining

https://www.sciencedirect.com/science/article/pii/S2949719123000213#b7

In this work we explored the relationship between discourse quality and memorization for LLMs. We found that the models that consistently output the highest-quality text are also the ones that have the highest memorization rate.

https://arxiv.org/abs/2601.02671

recent work shows that substantial amounts of copyrighted text can be extracted from open-weight models. However, it remains an open question if similar extraction is feasible for production LLMs, given the safety measures [...]. We investigate this question [...] our work highlights that, even with model- and system-level safeguards, extraction of (in-copyright) training data remains a risk for production LLMs.

How does merely tagging the apparently stolen content make it less problematic, given I'm guessing it still won't have any attribution of the actual source (which for all we know, might often even be GPL incompatible)?

But I'm not a lawyer, so I guess what do I know. But even from a non-legal angle, what is this road the Linux Foundation seems to embrace of just ignoring the license of projects? Why even have the kernel be GPL then, rather than CC0?

I don't get it. And the article calling this "pragmatism" seems absurd to me.

load more comments (30 replies)
[–] stylusmobilus@aussie.zone 25 points 1 week ago (3 children)

any resulting bugs or security flaws firmly onto the shoulders of the human submitting it.

Watch Americans and their companies pull some mad gymnastics on proportioning blame for this

load more comments (3 replies)
[–] pastermil@sh.itjust.works 19 points 1 week ago

Definitely not a big fan of it, but realistically speaking, it's here to stay. It is wise for them to govern and regulate it rather than outright ban it. Especially with a project as big as this one, people will try. Saying that the responsibility falls on the human is definitely the right move.

[–] twinnie@feddit.uk 13 points 1 week ago (10 children)

No point getting upset about this, it’s inevitable. So many FOSS programmers work thanklessly for hours and now there’s some tool to take loads of that work away, of course they’re going to use it. I know loads of people complain about it but used responsibly it can take care of so much of the mundane work. I used to spend 10% of my time writing code then 90% debugging it. If I do that 10% then give it to Claude to go over I find it just works.

load more comments (10 replies)
[–] XLE@piefed.social 12 points 1 week ago (7 children)

This seems like an ill-thought-out decision, especially in a landscape where Linux should be differentiating itself from, and not following Windows.

The titular "slop" just means "bad AI generated code is banned" but the definition of "bad" is as vague as Google's "don't be evil." Good luck enforcing it, especially in an open-source project where people's incentives aren't tied to a paycheck.

Title is also inaccurate regarding CoPilot (the Microsoft brand AI tool), as a comment there mentions

says yes to Copilot

Where in the article does it say that?? The only mention of CoPilot is where it talks about LLM-generated code having unverifiable provenance. Reply

[–] Naich@piefed.world 12 points 1 week ago

Google's "don't be evil" was like a warrant canary. It didn't need to be precise, it just needed to be there.

load more comments (6 replies)
load more comments
view more: next ›