this post was submitted on 07 May 2026
239 points (85.7% liked)

Technology

84478 readers
3729 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] Buffalox@lemmy.world 106 points 2 days ago* (last edited 2 days ago) (8 children)

According to a new study by researchers at Carnegie Mellon, MIT, Oxford, and UCLA,

Study should be solid I guess.

participants who were given AI assistants (in this case, a chatbot powered by OpenAI’s GPT-5 model) would have the aid pulled from them without warning during the test

Wow, interesting idea. 👍

where they had their assistant removed, the AI group saw the solve rate fall off a cliff. They had a solve rate about 20% lower

And even worse IMO:

They also had nearly double the skip rate, meaning they simply chose not to solve the questions.

This seems very alarming IMO, because this indicates they lost some of their ability to think constructively on how to actually solve a problem!

I know there have always been some who cried wold every time new technology has become available, like calculators and computers. Even dictionaries were once claimed to be harmful once!
But maybe this time there is a real danger, because AI takes away a lot of the need to actually think creatively and constructively. And that's an ability we must not lose.

The last paragraph of the article is even worse. As it mentions 2 studies that show these effects are also long term!!!

[–] Ioughttamow@fedia.io 55 points 2 days ago (1 children)

When driving somewhere, if I set out with the mindset that I can’t rely on gps I can usually wing it and figure out where to go when a hiccup occurs. If I don’t, then I have a lot of trouble getting into that path finding mode when needed… similar to this maybe?

[–] yakko@feddit.uk 20 points 2 days ago

Yeah exactly, because although it's possible to do more with technology sometimes, you're actively de-skilling at the same time. When we invented the written word yes it legitimately made everything better, but also we lost oral traditions and the capacity to memorize large volumes of storytelling, songs, and histories. Now you can burn the books, and the knowledge dies. It's a real risk.

Everything is like this. Every technology has a cost beyond its price, and making a decision of whether to use it or not will always be in error unless you think about what you're losing in the process.

[–] scarabic@lemmy.world 25 points 2 days ago (2 children)

Changing the terms of the test in the middle of it, without warning, is disruptive. I’m not convinced it “fried their brains.” The same would happen with a calculator suddenly removed during the middle of an exam.

[–] howrar@lemmy.ca 6 points 2 days ago

Or any task change really. You tell me that I'm here for a writing task, then halfway through it becomes a math test? There's no way I'm doing anywhere near as well as if they told me what was happening ahead of time.

[–] Buffalox@lemmy.world 3 points 2 days ago (1 children)

You are disregarding the last paragraph, where 2 other studies showed similar results, without having the "disruptive" factor.

[–] scarabic@lemmy.world 5 points 1 day ago* (last edited 1 day ago) (1 children)

Here’s that last paragraph. Microsoft’s finding actually sounds like it does have the disruptive factor: people are trained to use AI and then it is removed. And finally, finally in the very last sentence of the entire article we get the one piece of information that’s been missing the entire time: doctors perform better with AI help, but then worse than ever without it.

My conclusion? Let people have AI and perform better with it.

Carpenters trained on power tools will suddenly perform worse with hand tools than carpenters who were never given power tools. But if they are given power tools, they can build homes faster.

No shit?

The findings are also in line with a study Microsoft published last yearthat looked at cognitive decline among knowledge workers, which found that the more people lean on AI, the worse they perform when asked to work without support. It also echoes a study out of Poland, which found that while doctors are better at spotting cancer risks with AI assistance, they perform worse than the no-AI baseline once that assistance is removed.

[–] NeilNuggetstrong@lemmy.world 19 points 2 days ago (2 children)

If I use AI for my personal coding projects I've found that if the task is unsolvable by the ai model, I'm not able to sit down and do it myself until the next day. It's like I've got to reset my brain.

If I want to save time and use AI for a specific part of the code, it probably saves me 5 hours of work. But then I spend five hours yelling at the ai to try to get it to actually solve it. Next day I'll just fix it myself in 2 hours.

[–] sockenklaus@sh.itjust.works 5 points 21 hours ago* (last edited 21 hours ago) (1 children)

But what you're describing is not that uncommon, even without AI: Oftentimes when trying to solve a complex problem and being unsuccessful you have to reset your brain by doing something fundamentally different or have a good night of sleep and after that you solve the problem easily.

May what you're experiencing is not AI related at all.

[–] NeilNuggetstrong@lemmy.world 2 points 21 hours ago

You're probably right, but I think it's made worse by AI. Jumping into the code after 3 hours with Claude doing the dirty work feels like an impossibility

[–] Buffalox@lemmy.world 3 points 2 days ago (1 children)

That sounds a lot like what the studies show. And IMO that sounds like a serious problem.

[–] NeilNuggetstrong@lemmy.world 2 points 1 day ago

I'm really just tricking my brain to think I'm being more productive lmao.

But then again, some of the stuff I'm working on is in principle quite easy to do, but is also outside of my skillet, for these cases I benefit from using AI.

IMO the challenge is knowing how and when to use AI. Small companies using AI correctly can probably benefit massively from it. Although it's risky

[–] Tollana1234567@lemmy.today 0 points 1 day ago (1 children)

it has ruined the ability of K-12 people writing and reading proficiency.

[–] architect@thelemmy.club 2 points 20 hours ago

That was well in the toilet before llms.

[–] carotte@lemmy.blahaj.zone 10 points 2 days ago (2 children)

there have always been some who cried wold every time new technology has become available, like calculators and computers

and they kinda have a point, really. people got worse at memorizing stuff by heart when writing was invented, and people got worse at mental calculus when calculators when invented.

but they allowed many things that were simply not possible. a calculation that takes me 2 minutes in wolfram alpha could take hours if not days to solve by hand!

ai, meanwhile, or at least the ai we’re sold, does not offer significant advantages (at best it saves a few minutes), at the cost of making us worse at thinking, a skill that is absolutely essential to have… and of course, that’s the point. the tech oligarchs want us to be dependent on their extremely expensive products.

[–] Buffalox@lemmy.world 4 points 2 days ago (1 children)

and people got worse at mental calculus when calculators when invented.

That may be true, but that is a much more limited problem, than losing some of our ability for critical thinking and problem solving in general.

ai, meanwhile, or at least the ai we’re sold, does not offer significant advantages

This is very true, the AI are shown to even hallucinate, and give incorrect and harmful solutions. A calculator does NOT do that.
So not only is the AI a danger to our critical thinking, we actually need it MORE when using AI.

[–] architect@thelemmy.club 2 points 20 hours ago

A calculator will do that if you don’t know how to input correctly, and i promise you, plenty of people don’t know the rules of simple math.

[–] badgermurphy@lemmy.world 2 points 2 days ago

But they're using the hell out of it, too, right? They're exactly the types of people that love and use it the most: managers and owners.

[–] FauxLiving@lemmy.world 6 points 2 days ago (1 children)

This paper shows that a person who has performed a task 12 times performs better than a person who has never performed the same task.

They also do not properly control for performance loss due to context switching which is a well known contributor to performance loss.

It's a paper on arXiv, it hasn't been peer reviewed or published.

[–] Buffalox@lemmy.world 0 points 2 days ago (3 children)

No the test is not training, that's a weird thing to claim. The switch is what is tested, and you disregard that 2 other tests have shown similar results. An actual decline in critical and problem solving thinking.

[–] FauxLiving@lemmy.world 3 points 1 day ago* (last edited 1 day ago) (1 children)

Here is the paper: https://ai-project-website.github.io/AI-assistance-reduces-persistence/

No the test is not training, that’s a weird thing to claim.

The control group solved 12 questions manually and then the 3 test questions manually. The AI grouped solved 0 questions manually and the 3 test questions manually. One group had 12 more manual math tasks to prepare for the manual math test the other group had 0 and also had to context switch.

The AI-assisted group was dealt a context switch, which results in a pretty severe performance loss. A context switch causes performance loss of around 40% according to this paper, which was peer-reviewed and published and is also the most cited paper on the topic, in the APA: https://www.apa.org/pubs/journals/releases/xhp274763.pdf

The AI-assisted group also did not have 12 questions to adjust to the new context, like the control group did. If they wanted to wipe out the context switching performance loss they should have kept asking questions to see if, after 12 questions, the AI-assisted group had a similar performance.

The switch is what is tested, and you disregard that 2 other tests have shown similar results.

No, they did not switch what was tested. Here is an image from the actual paper.

They were given 12 tasks with one group using AI and another doing mental math and then 3 tasks doing mental math. One group had 12 more tasks worth of preparation than the other.

Nothing, not even the article in theOP, says that they did math and swapped to reading to test.

They did 3 different experiments, in each experiment they gave 12 tasks and then disabled the AI for one group and gave 3 more tasks as a test. At no point did they ask 12 math questions and then finish with 3 reading questions or vice versa. They did 2 experiments using math tasks and 1 experiment using reading comprehension tasks.

So one group had 15 math tasks and one group had 12 'how to ask an AI' tasks and then 3 math questions.

They also did not control for context switching losses, which is a well documented (see the APA paper) effect. The proper control would be to continue asking questions so the AI group also had 12 math tasks before the test.

There's a reason that this is published on arXiv and not in a peer-reviewed journal. Designing a poor quality experiment doesn't tell you anything useful even if you do multiple different versions of the same experiment.

This paper demonstrates a lack of a proper control group, specifically a failure to control for context switching performance loss.

[–] Buffalox@lemmy.world 0 points 1 day ago (2 children)

The picture you post contradict your claims. The 2 groups are getting the same question, but one has AI assistance, the other has not.
Again you fail to show anything to support your claims.

[–] sockenklaus@sh.itjust.works 2 points 15 hours ago (1 children)

No, what they meant is: The control group had 12 questions to get into the flow of solving math problems and then solved three more math problems for good measure.

The AI group on the other hand got into the flow of formulating math problems to ChatGPT and then had to actually solve three math problems themselves

Their critique is, that solving math problems yourself and prompting ChatGPT to solve math problems are not necessarily comparable tasks and require different skill sets so disabling AI after 12 tasks meant the first group had to switch context and therefore had worse performance.

If you want to analyze the first groups general ability of problem solving you should give them again twelve tasks after disabling AI so they get used to this new type of task (solving math problems yourself vs. prompting math problems to the AI) before measuring their performance.

[–] Buffalox@lemmy.world 1 points 15 hours ago

The AI group on the other hand got into the flow of formulating math problems to ChatGPT and then had to actually solve three math problems themselves

That's what the friggin test is about! So of course they did.

[–] FauxLiving@lemmy.world 4 points 23 hours ago (1 children)

I also wrote text.

If you're just going to cherry pick a single point and dismiss everything else then we're done here.

[–] frongt@lemmy.zip 3 points 20 hours ago (1 children)

Maybe they're unable to switch contexts

[–] FauxLiving@lemmy.world 2 points 19 hours ago

I hear that can cause a loss of performance.

[–] Womble@piefed.world 4 points 2 days ago* (last edited 2 days ago) (1 children)

The switch is what is being tested yes, but it is not clear that what is being measured in the switch is "AI fried their brains" rather than "context switching in the middle of a test". If they wanted to make that point it would be useful to have the maths test run with a calculator group who also got it yanked halfway through, that way we would be able to see what proportion of the effect is over dependence on AI removing critical thinking and what amount is having your methods disrupted mid task.

[–] Buffalox@lemmy.world 0 points 1 day ago

The calculator test might be good for comparison, and I'm pretty sure if given the same amount of time, and one group being allowed to use calculator for half the test, that group would solidly outperform a group not using calculators at all.

I was in 5th grade in 1975, and we were the first class to get calculators in 5th grade. Which became the standard for many years after.
I have never heard complaints about students being less capable of understanding basic math problems because they use calculators. Although the idea of using calculators in schools were heavily debated. It's similar to people not getting worse at spelling from using a dictionary.

[–] iglou@programming.dev 1 points 2 days ago

Not training, no, but warm up. And no, it is not about critical thinking, it's about reading comprehension and calculations.

[–] redsand@infosec.pub 5 points 2 days ago (2 children)

Also and this is the big one for me. It's 10% wrong on average. That's really bad. 1 in 10 google Gemini answers is bullshit

[–] Buffalox@lemmy.world 0 points 2 days ago

And the ability to think critically to detect it declines. So it's doubly harmful!

[–] ryannathans@aussie.zone 0 points 2 days ago

1 in 5 human answers is probably bullshit so it sounds like you're onto a winner

[–] derAbsender@piefed.social 0 points 2 days ago (1 children)

Wow. Now do this with a calculator.

[–] Buffalox@lemmy.world -1 points 2 days ago (2 children)

A calculator is not the same problem, it doesn't reduce our general ability to think critically.

[–] derAbsender@piefed.social 2 points 1 day ago (1 children)

As the study defines critical thinking: yes it does.

The study claims, essentially, relying on a machine that solves a Problem for you, lessens your critical thinking skills.

Their Definition of "critical thinking" is just, at least to me, way Off.

Just because i can conprehend Stuff i read for example, does not show critical thinking. It just shows i can repeat shit i read adequately.

It's just bad science.

[–] architect@thelemmy.club 1 points 20 hours ago

Oooooh have they done religion yet?

[–] iglou@programming.dev 2 points 2 days ago* (last edited 2 days ago) (1 children)

The studies referenced are about calculations, reading comprehension and work performance, not critical thinking.

The article is, like many, a bad one. It generalises what it should not.

[–] Buffalox@lemmy.world -1 points 1 day ago* (last edited 1 day ago) (1 children)

The sessions lasted about 10 minutes, suggesting that those who decided to rely heavily on AI to solve problems for them abandoned their critical thinking abilities in a matter of minutes.

[–] iglou@programming.dev 3 points 1 day ago* (last edited 1 day ago) (1 children)

As I said, this is a bad article. The experiment does not suggest that at all. The study does not mention critical thinking.

I'd say, however, that the proliferation of shitty news websites has caused readers to lose their critical thinking.

[–] Buffalox@lemmy.world -5 points 1 day ago* (last edited 1 day ago) (1 children)

In academia it is normal not to directly spell out things that are obvious to a person with academic knowledge on the subject, research papers are meant for scholars, and they are supposed to be able to read and understand the consequences for themselves.

So you can't use it as an argument that it isn't spelled out, if it can be easily derived by a person who understands the subject.
Research papers do not spell out every possible consequence of their findings.

[–] iglou@programming.dev 3 points 1 day ago* (last edited 1 day ago) (1 children)

It isn't spelled out because it is not a logical conclusion at all. Nothing in this test requires critical thinking to achieve.

Why are you defending an obviously terribly written article?

[–] architect@thelemmy.club 2 points 20 hours ago

Because AI bad!