Forwarding this invite to additional people who may wish to participate or
watch the recording.

Thanks, Kinneret. These presentations sound very interesting.

Pine🌲



---------- Forwarded message ---------
From: Kinneret Gordon via Analytics <[email protected]>
Date: Mon, Feb 23, 2026 at 5:23 AM
Subject: [Analytics] [Wikimedia Research Showcase] AI and Communities -
February 25 at 17:30 UTC
To: <[email protected]>, <
[email protected]>, <[email protected]>, <
[email protected]>
Cc: Kinneret Gordon <[email protected]>


Hi everyone,

The February 2026 Research Showcase will be live-streamed this Wednesday,
February 25, at 9:30 AM PT / 17:30 UTC. Find your local time here
<https://zonestamp.toolforge.org/1772040600>. Our theme this month is *AI
and Communities*.

*We invite you to watch via the YouTube
stream: https://www.youtube.com/live/qW5IQJv84HY
<https://www.youtube.com/live/qW5IQJv84HY>.* As always, you can join the
conversation in the YouTube chat as soon as the showcase goes live.

This month, we will have two presentations:

*LLMs in Wikipedia: Investigating How LLMs Impact Participation in
Knowledge Communities*By *Moyan Zhou (University of Minnesota)*Large
language models (LLMs) are reshaping knowledge production as community
members increasingly incorporate them into their contribution workflows.
However, participating in knowledge communities involves more than just
contributing content - it is also a deeply social process shaped by
members' level of expertise. While communities must carefully consider
appropriate and responsible LLM integration, the absence of concrete norms
has left individual editors to experiment and navigate LLM use on their
own. Understanding how LLMs influence community participation across
expertise levels is therefore critical in shaping future norms and
supporting effective adoption. To address this gap, we investigated
Wikipedia, one of the largest knowledge production communities, to
understand participation in three dimensions: 1) how LLMs influence the
ways editors gather knowledge, 2) how editors leverage strategies to align
LLM outputs with community norms, and 3) how other editors in the community
respond to LLM-assisted contributions. Through interviews with 16 Wikipedia
editors with different levels of expertise who had used LLMs for their
edits, we revealed a participation gap mediated by expertise in adopting
LLMs in knowledge contributions across knowledge gathering, alignment with
community norms, and peer responses. Based on these findings, we challenge
existing models of novice editors' involvement and propose design
implications for LLMs that support community engagement, highlighting
opportunities for LLMs to sustain mentorship, knowledge transmission, and
legitimacy building by scaffolding and feedback, process documentation, and
LLM disclosure by good-faith editors.*AI Didn't Start the Fire: Examining
the Stack Exchange Moderator and Contributor Strike*By *Yiwei Wu
(University of Texas at Austin)*Online communities and their host platforms
are mutually dependent yet conflict-prone. When platform policies clash
with community values, communities have resisted through strikes,
blackouts, and even migration to other platforms. Through such collective
actions, communities have sometimes won concessions, but these have
frequently proved to be temporary. Although previous research has
investigated strike events and migration chains, the processes by which
community-platform conflict unfolds remain obscure. How do
community-platform relationships deteriorate? How do communities organize
collective action? How do the participants proceed in the aftermath? We
investigate a conflict between the Stack Exchange platform and community
that occurred in 2023 around an emergency arising from the release of large
language models (LLMs). Based on a qualitative thematic analysis of 2,070
messages from Meta Stack Exchange and 14 interviews with community members,
we reveal how the 2023 conflict was preceded by a long-term deterioration
in the community-platform relationship, driven in particular by the
platform's disregard for the community's highly valued participatory role
in governance. Moreover, the platform's policy response to LLMs aggravated
the community's sense of crisis, triggering strike mobilization. We analyze
how the mobilization was coordinated through a tiered leadership and
communication structure, as well as how community members pivoted in the
aftermath. Building on recent theoretical scholarship in social computing,
we use Hirschman's exit, voice, and loyalty framework to theorize the
challenges of community-platform relations evinced in our data. Finally, we
recommend ways that platforms and communities can institute participatory
governance to be durable and effective.
Looking forward to seeing many of you,
Kinneret
-- 

Kinneret Gordon

Lead Research Community Officer

Wikimedia Foundation <https://wikimediafoundation.org/>

*Learn more about Wikimedia Research <https://research.wikimedia.org/>*
_______________________________________________
Analytics mailing list -- [email protected]
To unsubscribe send an email to [email protected]
_______________________________________________
Wikitech-l mailing list -- [email protected]
To unsubscribe send an email to [email protected]
https://lists.wikimedia.org/postorius/lists/wikitech-l.lists.wikimedia.org/

Reply via email to