Dear colleagues,

We are delighted to announce SemEval-2026 Task 3 Track B: Dimensional
Stance Analysis

*Aspect-Based Sentiment Analysis (ABSA)* is a widely used technique for
analyzing people’s opinions and sentiments at the aspect level. However,
current ABSA research predominantly adopts a coarse-grained, categorical
sentiment representation (e.g., positive, negative, or neutral). This
approach stands in contrast to long-established theories in psychology and
affective science, where sentiment is represented along fine-grained,
real-valued dimensions of valence (ranging from negative to positive) and
arousal (from sluggish to excited). This valence-arousal (VA)
representation has inspired the rise of dimensional sentiment analysis as
an emerging research paradigm, enabling more nuanced distinctions in
emotional expression and supporting a broader range of applications.

Given an utterance or post and a target entity, stance detection involves
determining whether the speaker is in favor or against the target. *This
track reformulates stance detection as a Stance-as-DimABSA task with the
following transformations:*


*1. The stance target is treated as an aspect.2. Discrete stance labels are
replaced with continuous VA scores.*

Building on this, we introduce *Dimensional Stance Analysis (DimStance)*, a
Stance-as-DimABSA task that reformulates stance detection under the ABSA
schema in the VA space. This new formulation extends ABSA beyond consumer
reviews to public-issue discourse (i.e., politics and environmental
protection) and also generalizes stance analysis from categorical labels to
continuous VA scores. Given a text and one or more aspects (targets),
predict a real-valued valence-arousal (VA) score for each aspect,
reflecting the stance expressed by the speaker toward it.

———————
*Languages*
———————
*We provide data in 5 languages*, including: German (deu), English (eng),
Hausa (hau), Swahili (swa), and Chinese (zho)

———————
*Evaluation*
———————
RMSE is used.

———————
*Participation*
———————
*Website* (checkout details):
https://github.com/DimABSA/DimABSA2026

*Codabench* (register and submit results)
- Track B: https://www.codabench.org/competitions/11139/

*Discord* (community and discussion)
https://discord.gg/xWXDWtkMzu

*Google Group* (official updates):
https://groups.google.com/g/dimabsa-participants

———————
*Important Dates *
———————
- Sample Data Ready: 15 July 2025
- Training Data Ready: 30 September 2025
- Evaluation Start: 12 January 2026
- Evaluation End: 30 January 2026
- System Description Paper Due: February 2026
- Notification to Authors: March 2026
- Camera Ready Due: April 2026
- SemEval Workshop 2026: co-located with ACL 2026 (San Diego, CA, USA)

We warmly invite the community to participate in this exciting shared task
and contribute to advancing NLP research.

Best regards,
SemEval-2026 Task 3 Organizers
_______________________________________________
Corpora mailing list -- [email protected]
https://list.elra.info/mailman3/postorius/lists/corpora.list.elra.info/
To unsubscribe send an email to [email protected]

Reply via email to