this post was submitted on 08 Jun 2025
774 points (95.7% liked)

Technology

71146 readers
2860 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
 

LOOK MAA I AM ON FRONT PAGE

you are viewing a single comment's thread
view the rest of the comments
[–] Knock_Knock_Lemmy_In@lemmy.world 3 points 8 hours ago (1 children)

do we know that they don't and are incapable of reasoning.

"even when we provide the algorithm in the prompt—so that the model only needs to execute the prescribed steps—performance does not improve"

[–] communist@lemmy.frozeninferno.xyz 1 points 7 hours ago* (last edited 6 hours ago) (1 children)

That indicates that this particular model does not follow instructions, not that it is architecturally fundamentally incapable.

[–] Knock_Knock_Lemmy_In@lemmy.world 2 points 6 hours ago (1 children)

Not "This particular model". Frontier LRMs s OpenAI’s o1/o3,DeepSeek-R, Claude 3.7 Sonnet Thinking, and Gemini Thinking.

The paper shows that Large Reasoning Models as defined today cannot interpret instructions. Their architecture does not allow it.

[–] communist@lemmy.frozeninferno.xyz 1 points 5 hours ago* (last edited 5 hours ago) (2 children)

those particular models. It does not prove the architecture doesn't allow it at all. It's still possible that this is solvable with a different training technique, and none of those are using the right one. that's what they need to prove wrong.

this proves the issue is widespread, not fundamental.

[–] 0ops@lemm.ee 3 points 4 hours ago (1 children)

Is "model" not defined as architecture+weights? Those models certainly don't share the same architecture. I might just be confused about your point though

[–] communist@lemmy.frozeninferno.xyz 2 points 4 hours ago* (last edited 3 hours ago) (1 children)

It is, but this did not prove all architectures cannot reason, nor did it prove that all sets of weights cannot reason.

essentially they did not prove the issue is fundamental. And they have a pretty similar architecture, they're all transformers trained in a similar way. I would not say they have different architectures.

[–] 0ops@lemm.ee 1 points 3 hours ago
[–] Knock_Knock_Lemmy_In@lemmy.world 1 points 3 hours ago (1 children)

The architecture of these LRMs may make monkeys fly out of my butt. It hasn't been proven that the architecture doesn't allow it.

You are asking to prove a negative. The onus is to show that the architecture can reason. Not to prove that it can't.

[–] communist@lemmy.frozeninferno.xyz 2 points 1 hour ago* (last edited 1 hour ago)

that's very true, I'm just saying this paper did not eliminate the possibility and is thus not as significant as it sounds. If they had accomplished that, the bubble would collapse, this will not meaningfully change anything, however.

also, it's not as unreasonable as that because these are automatically assembled bundles of simulated neurons.