this post was submitted on 12 Jun 2025
117 points (100.0% liked)

Technology

39201 readers
349 users here now

A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.

Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.

Subcommunities on Beehaw:


This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

founded 3 years ago
MODERATORS
 

“The Wikimedia Foundation has been exploring ways to make Wikipedia and other Wikimedia projects more accessible to readers globally,” a Wikimedia Foundation spokesperson told me in an email. “This two-week, opt-in experiment was focused on making complex Wikipedia articles more accessible to people with different reading levels. For the purposes of this experiment, the summaries were generated by an open-weight Aya model by Cohere. It was meant to gauge interest in a feature like this, and to help us think about the right kind of community moderation systems to ensure humans remain central to deciding what information is shown on Wikipedia.”

Some very out of touch people in the Wikimedia Foundation. Fortunately the editors (people who actually write the articles) have the sense to oppose this move in mass.

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 5 points 5 days ago (5 children)

Wikipedia would probably be the only organization that I would trust with AI. They've been using it for a while now to flag sections that might need to be rewritten, but they don't let the AI write anything itself, only notify human editors that there might be a problem. Or, at least that was what I heard a couple of ywars ago when they talked about it last.

[–] [email protected] 15 points 5 days ago (1 children)

That is not the case here. These are not bots which flagged issues, but literally a LLM to help with writing "summaries", which is why the reaction is so different.

[–] [email protected] 4 points 5 days ago

Yeah, I was thinking that if any organization would do AI summaries right, it would be Wikipedia. But I trust the editors the most.

load more comments (3 replies)