Call for Papers

1st Workshop on Reliable Evaluation of LLMs for Factual Information (REAL-Info)

Co-located with ICWSM 2024, June 3, 2024, Buffalo, NY

https://sites.google.com/view/real-info-2024

LLMs have achieved state-of-the-art performance in several textual inference 
tasks and are gaining popularity. There is a significant focus on their 
integration with web and online applications, including web search, thus 
allowing them to reach millions of users. LLMs can influence various 
information tasks in our everyday lives, ranging from personal content creation 
to education, financial advice, and mental health support (Augenstein, 2023). 
However, with their vast linguistic capabilities and opaque nature, LLMs can 
inadvertently generate or amplify false information. There is growing concern 
about the factuality of LLM-generated content and its potential adverse impact 
on our information ecosystem (Chen, 2023; Peskoff, 2023).

Thus the need for reliable methods to assess the factuality of information is 
more critical than ever. This is where the synergy of AI, Natural Language 
Processing (NLP), and Human-Computer Interaction (HCI) becomes essential. AI 
and NLP techniques can be employed to analyze and identify the factuality of 
information through various tasks (Augenstein, 2023), such as fact-checking, 
stance detection, claim verification, and misinformation detection. These 
techniques can sift through the vast amounts of data to spot inconsistencies, 
biases, or inaccuracies that could indicate misinformation. Still, these 
approaches often use language models themselves, and epistemological questions 
arise when one LLM is fact-checked using another (or itself). Meanwhile, HCI 
plays a vital role in designing interactions and tools that enable humans to 
effectively oversee, interpret, and correct the outputs of LLMs. This 
human-in-the-loop approach ensures a critical evaluation and context-sensitive 
understanding of the factuality of information, which pure algorithmic methods 
might overlook. The combination of NLP's analytical capabilities and HCI's 
focus on human-centric design is instrumental in creating a digital ecosystem 
where LLMs can be utilized safely and responsibly, minimizing the risks of 
false information while maximizing their potential for user-centric 
applications.

The goals of the 1st ICWSM workshop Reliable Evaluation of LLMs for Factual 
Information (REAL-Info) are to facilitate discussion around such new LLM 
evaluation approaches, metrics, and benchmarks for factuality assessment tasks 
within the community, to inform the scope, biases, and blindspots of LLMs. It 
will spark interdisciplinary conversations from academic and industry 
researchers in computational social sciences (CSS), natural language processing 
(NLP), human-computer interaction (HCI), data science, and social computing. 
The workshop will solicit, research, and position papers with novel ideas, 
including but not limited to:
- New evaluation methods and metrics for evaluating LLM’s factuality 
considering diverse social context, e.g., source and domain of data, language, 
temporal generalization of information, or hallucination in 
generated/summarized content.
- Human-centered design approaches to aid LLMs in detecting and mitigating 
false information, e.g., human experts in the loop, and variation in prompting.
- New LLM-powered tools, methods, and applications for improving factuality 
assessment in social computing and computational social science.
- Biases and blindspots of LLMs in factuality assessment, including approaches 
for error analysis and model diagnostics.
- Limitations of existing benchmarks for tasks relevant to factuality 
assessment, e.g., claim verification, fact-checking, stance detection, and 
misinformation detection.
- Improve datasets and evaluation quality, e.g., avoidance of selection bias, 
addressing subjective judgments and biases in crowd-sourced annotation.
- Comparative evaluation and implications of open source and commercial LLMs 
for tasks relevant to factuality assessment.
- How does the reliability and factuality of LLM impact users (e.g. 
journalists, software engineers, artists..) and communities?

Submission instructions can be found on the workshop website. The workshop will 
take place as a half-day meeting in June. Authors of accepted papers will have 
the opportunity to publish their papers through workshop proceedings by the 
AAAI Press.

Timeline
- Workshop Papers Submission deadline: March 24, 2024
- Notifications: April 14, 2024
- Final Camera-Ready Paper Due: May 5, 2024
- ICWSM-2024 Workshops Day: June 3, 2024
The University of Edinburgh is a charitable body, registered in Scotland, with 
registration number SC005336. Is e buidheann carthannais a th’ ann an Oilthigh 
Dhùn Èideann, clàraichte an Alba, àireamh clàraidh SC005336.
_______________________________________________
Corpora mailing list -- [email protected]
https://list.elra.info/mailman3/postorius/lists/corpora.list.elra.info/
To unsubscribe send an email to [email protected]

Reply via email to