this post was submitted on 16 Jun 2025
98 points (98.0% liked)

Technology

71502 readers
4727 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
 

geteilt von: https://feddit.org/post/14273969

This is a paper for a MIT study. Three groups of participants where tasked to write an essay. One of them was allowed to use a LLM. These where the results:

The participants mental activity was also checked repeatedly via EEG. As per the papers abstract:

EEG revealed significant differences in brain connectivity: Brain-only participants exhibited the strongest, most distributed networks; Search Engine users showed moderate engagement; and LLM users displayed the weakest connectivity. Cognitive activity scaled down in relation to external tool use.

top 4 comments
sorted by: hot top controversial new old
[–] [email protected] 14 points 1 day ago

Shows people using the LLM thought less about it and probably learned less.

Sometimes students are motivated to learn, but there's always the temptation to just get the work done and relax...

[–] [email protected] 13 points 1 day ago

It's like using a faulty calculator, but with even less mental investment.

[–] [email protected] 4 points 1 day ago

This is interesting to me as I like to say the llms are basically another abstraction of search. Initially it was links with no real weight that had to be gone through and then various algorithms weighted the return, then the results started giving a small blurb so one did not have to follow every link, and now your basically getting a report which should have references to the sources. I would like to see this looking at how folks engage with an llm. Basically my guess is if one treats the llm as a helper and collaborates to create the product that they will remember more than if they treat it as a servant and just instructs them to do it and takes the output as is.

[–] limer -5 points 1 day ago

I bet this is just correlation. Frequent reliance on AI ( and not just using it as a tool for a specific reason)