this post was submitted on 16 Jun 2025
66 points (100.0% liked)
Fuck AI
3139 readers
1054 users here now
"We did it, Patrick! We made a technological breakthrough!"
A place for all those who loathe AI to discuss things, post articles, and ridicule the AI hype. Proud supporter of working people. And proud booer of SXSW 2024.
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
The biggest flaw in this study is that the LLM group wasn’t ~~allowed~~ explicitly permitted to edit their essays and was explicitly forbidden from altering the parameters. Of course brain activity looks low if you just copy-paste a bot’s output without thinking. That’s not "using a tool"; that’s outsourcing cognition.
If you don’t bother to review, iterate, or humanize the AI’s output, then yeah... it’s a self-fulfilling prophecy: no thinking in, no thinking out.
In any real academic setting, “fire-and-forget” turns into “fuck around and find out” pretty quick.
LLMs aren’t the problem; they’re tools. Even journal authors use them. Blaming the tech instead of the lazy-ass operator is like saying:
No, he’s just using better tools. The problem is if he can’t build a chair afterward.
I didn't read the whole thing but only skimmed through the protocol. I only spotted
"participants were instructed to pick a topic among the proposed prompts, and then to produce an essay based on the topic's assignment within a 20 minutes time limit. Depending on the participant's group assignment, the participants received additional instructions to follow: those in the LLM group (Group 1) were restricted to using only ChatGPT, and explicitly prohibited from visiting any websites or other LLM bots. The ChatGPT account was provided to them. They were instructed not to change any settings or delete any conversations."
which I don't interpret as no editing. Can you please share where you found that out?
Lol, oops, I got poo brain right now. I inferred they couldn't edit because the methodology doesn't say whether revisions were allowed.
What is clear, is they weren't permitted to edit the prompt or add personalization details seems to imply the researchers weren't interested in understanding how a participant might use it in a real setting; just passive output. This alone undermines the premise.
It makes it hard to assess whether the observed cognitive deficiency was due to LLM assistance, or the method by which it was applied.
The extent of our understanding of the methodology is that they couldn't delete chats. If participants were only permitted to a a one-shot generation per prompt, then there's something wrong.
But just as concerning is the fact that it isnt explicitly stated.