Hello everyone,

This is a friendly reminder that this month's research showcase on *Ensuring
Content Integrity on Wikipedia *will be starting in an hour at 9:30 AM PT /
16:30 UTC. *We invite you to watch via the YouTube
stream: https://www.youtube.com/live/GgYh6zbrrss
<https://www.youtube.com/live/GgYh6zbrrss>.*

Best,
Kinneret

On Thu, Jun 12, 2025 at 8:12 PM Kinneret Gordon <kgor...@wikimedia.org>
wrote:

> Hi everyone,
>
> The June 2025 Research Showcase will be live-streamed next Wednesday, June
> 18, at 9:30 AM PT / 16:30 UTC. Find your local time here
> <https://zonestamp.toolforge.org/1750264200>. Our theme this month is 
> *Ensuring
> Content Integrity on Wikipedia*.
>
> *We invite you to watch via the YouTube
> stream: https://www.youtube.com/live/GgYh6zbrrss
> <https://www.youtube.com/live/GgYh6zbrrss>.* As always, you can join the
> conversation in the YouTube chat as soon as the showcase goes live.
>
> Our presentations this month:
> The Differential Effects of Page Protection on Wikipedia Article Quality
> By
> *Manoel Horta Ribeiro (Princeton University)*Wikipedia strives to be an
> open platform where anyone can contribute, but that openness can sometimes
> lead to conflicts or coordinated attempts to undermine article quality. To
> address this, administrators use “page protection"—a tool that restricts
> who can edit certain pages. But does this help the encyclopedia, or does it
> do more harm than good? In this talk, I’ll present findings from a
> large-scale, quasi-experimental study using over a decade of English
> Wikipedia data. We focus on situations where editors requested page
> protection and compare the outcomes for articles that were protected versus
> similar ones that weren’t. Our results show that page protection has mixed
> effects: it tends to benefit high-quality articles by preventing decline,
> but it can hinder improvement in lower-quality ones. These insights reveal
> how protection shapes Wikipedia content and help inform when it’s most
> appropriate to restrict editing, and when it might be better to leave the
> page open.
>
> Seeing Like an AI: How LLMs Apply (and Misapply) Wikipedia Neutrality Norms
> By
> *Joshua Ashkinaze (University of Michigan)*Large language models (LLMs)
> are trained on broad corpora and then used in communities with specialized
> norms. Is providing LLMs with community rules enough for models to follow
> these norms? We evaluate LLMs' capacity to detect (Task 1) and correct
> (Task 2) biased Wikipedia edits according to Wikipedia's Neutral Point of
> View (NPOV) policy. LLMs struggled with bias detection, achieving only 64%
> accuracy on a balanced dataset. Models exhibited contrasting biases (some
> under- and others over-predicted bias), suggesting distinct priors about
> neutrality. LLMs performed better at generation, removing 79% of words
> removed by Wikipedia editors. However, LLMs made additional changes beyond
> Wikipedia editors' simpler neutralizations, resulting in high-recall but
> low-precision editing. Interestingly, crowdworkers rated AI rewrites as
> more neutral (70%) and fluent (61%) than Wikipedia-editor rewrites.
> Qualitative analysis found LLMs sometimes applied NPOV more comprehensively
> than Wikipedia editors but often made extraneous non-NPOV-related changes
> (such as grammar). LLMs may apply rules in ways that resonate with the
> public but diverge from community experts. While potentially effective for
> generation, LLMs may reduce editor agency and increase moderation workload
> (e.g., verifying additions). Even when rules are easy to articulate, having
> LLMs apply them like community members may still be difficult.
>
> Best,
> Kinneret
>
> --
>
> Kinneret Gordon
>
> Lead Research Community Officer
>
> Wikimedia Foundation <https://wikimediafoundation.org/>
>
> *Learn more about Wikimedia Research <https://research.wikimedia.org/>*
>
_______________________________________________
Wiki-research-l mailing list -- wiki-research-l@lists.wikimedia.org
To unsubscribe send an email to wiki-research-l-le...@lists.wikimedia.org

Reply via email to