this post was submitted on 27 Apr 2026
1169 points (97.7% liked)

Science Memes

20042 readers
1908 users here now

Welcome to c/science_memes @ Mander.xyz!

A place for majestic STEMLORD peacocking, as well as memes about the realities of working in a lab.



Rules

  1. Don't throw mud. Behave like an intellectual and remember the human.
  2. Keep it rooted (on topic).
  3. No spam.
  4. Infographics welcome, get schooled.

This is a science community. We use the Dawkins definition of meme.



Research Committee

Other Mander Communities

Science and Research

Biology and Life Sciences

Physical Sciences

Humanities and Social Sciences

Practical and Applied Sciences

Memes

Miscellaneous

founded 3 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] Deebster@programming.dev 59 points 1 day ago (3 children)

Have they taken out the AI generated papers? We know that training LLMs on LLM-generated text leads to an absolute collapse in quality, and we also know that AI has been showing up in papers so if they haven't, then this will be quite unreliable.

[–] brucethemoose@lemmy.world 36 points 1 day ago* (last edited 1 day ago)

We know that training LLMs on LLM-generated text leads to an absolute collapse in quality.

This is often repeated, and true. But needs to be qualified.

Modern LLMs use tons and tons of “augmented” data, which is code for LLM generated or massaged data. Some is even generated during training, and judged; papers on that are what made Deepseek famous.

Training on LLM trash will, of course, yield greater trash, and obviously good text has to come from something real. But that’s because slop is slop. And there are issues with “deep frying” LLMs, yes, but simply training on LLM on LLM output does not necessarily reduce quality. It often helps, significantly.


And we also know that AI has been showing up in papers so if they haven’t, then this will be quite unreliable.

Now this is a problem.

TBH LLMs would be pretty good at flagging papers for humans to check, similar to what Wikipedia is already doing. But yeah, if you just feed a prompt bad papers, LLMs just assume the context is true, generally, and that’s a tremendous problem.

[–] T156@lemmy.world 7 points 1 day ago (1 children)

I would be surprised if it was something that they trained themselves, and not an off the shelf model hooked up to a search.

[–] brucethemoose@lemmy.world 1 points 1 day ago* (last edited 1 day ago)

It's probably their own search/RAG backend, or at least their configuration of some open source project.

And that's the important part. Get the article retrieval right, and the LLM performance isn't that important; they could self-host Qwen 27B or something and it'd work fine.