68
submitted 10 months ago* (last edited 10 months ago) by Vittelius@feddit.org to c/fuck_ai@lemmy.world

This is a paper for a MIT study. Three groups of participants where tasked to write an essay. One of them was allowed to use a LLM. These where the results:

The participants mental activity was also checked repeatedly via EEG. As per the papers abstract:

EEG revealed significant differences in brain connectivity: Brain-only participants exhibited the strongest, most distributed networks; Search Engine users showed moderate engagement; and LLM users displayed the weakest connectivity. Cognitive activity scaled down in relation to external tool use.

all 9 comments
sorted by: hot top new old
[-] sp3ctr4l@lemmy.dbzer0.com 11 points 10 months ago* (last edited 10 months ago)

Oh neat, a new subclass of brainrot has been observed and defined.

I've been saying this for years:

One way to win the Turing test is if you make every human so much more stupid that it is easier for an 'AI' to pass the test.

Now that's out of the box thinking!

Unfortunately it is a massive negative sum solution to a game theoretic approach to passing a Turing test, but it is a solution.

Roughly on par with winning an arcade fighting game by actually punching your local opponent irl.

[-] utopiah@lemmy.world 5 points 10 months ago* (last edited 10 months ago)

Very interesting, emphasis mine :

"findings support the view that external support tools restructure not only task performance but also the underlying cognitive architecture. The Brain-only group leveraged broad, distributed neural networks for internally generated content; the Search Engine group relied on hybrid strategies of visual information management and regulatory control; and the LLM group optimized for procedural integration of AI-generated suggestions.

These distinctions carry significant implications for cognitive load theory, the extended mind hypothesis [102], and educational practice. As reliance on AI tools increases, careful attention must be paid to how such systems affect neurocognitive development, especially the potential trade-offs between external support and internal synthesis."

Also the focus on agency and ownership is also very interesting, namely regardless of the scored outcome or how one might think the work itself changed them, or not, do they themselves feel it is their work?

[-] technocrit@lemmy.dbzer0.com 1 points 10 months ago* (last edited 10 months ago)

do they themselves feel it is their work?

Why should they care? They've gone into this artificial situation for money and they're just copying/writing words to get paid. It's similar to school in that sense. Pointless work in order to please people in power.

[-] Zozano@aussie.zone 1 points 10 months ago* (last edited 10 months ago)

The biggest flaw in this study is that the LLM group wasn’t ~~allowed~~ explicitly permitted to edit their essays and was explicitly forbidden from altering the parameters. Of course brain activity looks low if you just copy-paste a bot’s output without thinking. That’s not "using a tool"; that’s outsourcing cognition.

If you don’t bother to review, iterate, or humanize the AI’s output, then yeah... it’s a self-fulfilling prophecy: no thinking in, no thinking out.

In any real academic setting, “fire-and-forget” turns into “fuck around and find out” pretty quick.

LLMs aren’t the problem; they’re tools. Even journal authors use them. Blaming the tech instead of the lazy-ass operator is like saying:

These people got swole by hand-sawing wood, but this pudgy fucker used a power saw to cut 20 pieces faster; clearly he’s doing it wrong.

No, he’s just using better tools. The problem is if he can’t build a chair afterward.

[-] utopiah@lemmy.world 4 points 10 months ago* (last edited 10 months ago)

The biggest flaw in this study is that the LLM group wasn’t allowed to edit their essays

I didn't read the whole thing but only skimmed through the protocol. I only spotted

"participants were instructed to pick a topic among the proposed prompts, and then to produce an essay based on the topic's assignment within a 20 minutes time limit. Depending on the participant's group assignment, the participants received additional instructions to follow: those in the LLM group (Group 1) were restricted to using only ChatGPT, and explicitly prohibited from visiting any websites or other LLM bots. The ChatGPT account was provided to them. They were instructed not to change any settings or delete any conversations."

which I don't interpret as no editing. Can you please share where you found that out?

The biggest flaw in this study is that the LLM group wasn’t allowed to edit their essays

[-] Zozano@aussie.zone -3 points 10 months ago* (last edited 10 months ago)

Lol, oops, I got poo brain right now. I inferred they couldn't edit because the methodology doesn't say whether revisions were allowed.

What is clear, is they weren't permitted to edit the prompt or add personalization details seems to imply the researchers weren't interested in understanding how a participant might use it in a real setting; just passive output. This alone undermines the premise.

It makes it hard to assess whether the observed cognitive deficiency was due to LLM assistance, or the method by which it was applied.

The extent of our understanding of the methodology is that they couldn't delete chats. If participants were only permitted to a a one-shot generation per prompt, then there's something wrong.

But just as concerning is the fact that it isnt explicitly stated.

[-] technocrit@lemmy.dbzer0.com 1 points 10 months ago* (last edited 10 months ago)

Maybe the actual problem is sorting and filtering people based on pointless essays. It selects for the most privileged and obedient.

this post was submitted on 16 Jun 2025
68 points (100.0% liked)

Fuck AI

6789 readers
1072 users here now

"We did it, Patrick! We made a technological breakthrough!"

A place for all those who loathe AI to discuss things, post articles, and ridicule the AI hype. Proud supporter of working people. And proud booer of SXSW 2024.

AI, in this case, refers to LLMs, GPT technology, and anything listed as "AI" meant to increase market valuations.

founded 2 years ago
MODERATORS