- cross-posted to:
- technology@lemmy.world
- cross-posted to:
- technology@lemmy.world
“The Wikimedia Foundation has been exploring ways to make Wikipedia and other Wikimedia projects more accessible to readers globally,” a Wikimedia Foundation spokesperson told me in an email. “This two-week, opt-in experiment was focused on making complex Wikipedia articles more accessible to people with different reading levels. For the purposes of this experiment, the summaries were generated by an open-weight Aya model by Cohere. It was meant to gauge interest in a feature like this, and to help us think about the right kind of community moderation systems to ensure humans remain central to deciding what information is shown on Wikipedia.”
Some very out of touch people in the Wikimedia Foundation. Fortunately the editors (people who actually write the articles) have the sense to oppose this move in mass.
Wikipedia would probably be the only organization that I would trust with AI. They’ve been using it for a while now to flag sections that might need to be rewritten, but they don’t let the AI write anything itself, only notify human editors that there might be a problem. Or, at least that was what I heard a couple of ywars ago when they talked about it last.
That is not the case here. These are not bots which flagged issues, but literally a LLM to help with writing “summaries”, which is why the reaction is so different.
Yeah, I was thinking that if any organization would do AI summaries right, it would be Wikipedia. But I trust the editors the most.
For som reason, “ywars” changed your voice into that of a pirate, and it made me cackle. Thanks 💛
Lmao. Yeah, I don’t use auto-complete and rarely re-read things when I write, so mistakes are bound to happen :P
Fair. I should really quit using autocomplete and stop using Gboard for privacy reasons. Honestly, I’m just a little bit away from de-googling and going graphene. Just gotta spin up immich and a few other servers.