Dear Sender,

I am currently out of the office and will not be checking emails regularly. I 
will return on September 9, and will respond to your message as soon as 
possible after that date.

Best regards,
Charlott Jakob

On 17 Jul 2024, at 18:47, Verna via Corpora <[email protected]> wrote:

GenBench: The second workshop on generalisation (benchmarking) in NLP

Workshop description
The ability to generalise well is often mentioned as one of the primary 
desiderata for models of natural language processing (NLP).
Yet, there are still many open questions related to what it means for an NLP 
model to generalise well, and how generalisation should be evaluated.
LLMs, trained on gigantic training corpora that are – at best – hard to analyse 
or not publicly available at all, bring a new set of challenges to the topic.
The second GenBench workshop aims to serve as a cornerstone to catalyse 
research on generalisation in the NLP community.
The workshop aims to bring together different expert communities to discuss 
challenging questions relating to generalisation in NLP, crowd-source 
challenging generalisation benchmarks for LLMs, and make progress on open 
questions related to generalisation.

Topics of interest include, but are not limited to:
        •       Opinion or position papers about generalisation and how it 
should be evaluated;
        •       Analyses of how existing or new models generalise;
        •       Empirical studies that propose new paradigms to evaluate 
generalisation;
        •       Meta-analyses that compare how results from different 
generalisation studies compare;
        •       Meta-analyses that study how different types of generalisation 
are related;
        •       Papers that discuss how generalisation of LLMs can be evaluated;
        •       Papers that discuss why generalisation is (not) important in 
the era of LLMs;
        •       Studies on the relationship between generalisation and fairness 
or robustness.
The second GenBench workshop on generalisation (benchmarking) in NLP will be 
co-located with EMNLP 2024.

Submission types
We call for two types of submissions: regular workshop submissions and 
collaborative benchmarking task submissions.
The latter will consist of a data/task artefact and a companion paper 
motivating and evaluating the submission.
In both cases, we accept archival papers and extended abstracts.

1. Regular workshop submissions
Regular workshop submissions present papers on the topic of generalisation (see 
examples listed above).
Regular workshop papers may be submitted as an archival paper, when they report 
on completed, original and unpublished research, or as a shorter extended 
abstract, otherwise.
More details on this category can be found below.
If you are unsure whether a specific topic is well-suited for submission, feel 
free to reach out to the organisers of the workshop at 
[email protected].

2. Collaborative Benchmarking Task (CBT) submissions
The goal of this year's CBT is to generate versions of existing evaluation 
datasets for LLMs which, given a particular training corpus, have a larger 
distribution shift than the original test set, or – in other words – evaluate 
generalisation to a stronger degree than the original dataset.
For this particular challenge, we focus on three training corpora: C4, 
RedPajama-Data-1T, and Dolma.
All three corpora are publicly available, and they can be searched via the 
What's in My Big Data API (https://github.com/allenai/wimbd).
We will focus on three popular evaluation datasets: MMLU, HumanEval, and SiQA.
Submitters to the CBT are asked to design a way to assess distribution shift 
for one or more of these evaluation datasets, given particular features of the 
training corpus, and then generate one or more versions of the dataset that 
have a larger distribution shift according to this method.
Newly generated sets do not have to have the same size as the original test 
set, but should have at least 200 examples.
Practically speaking, CBT submissions consist of:
        1.      the data/task artefact, submitted through 
https://github.com/GenBench/genbench_cbt
        2.      a paper describing the dataset and its method of construction, 
submitted through https://openreview.net/group?id=GenBench.org/2024/Workshop
We accept submissions that consider only one pretraining dataset and evaluation 
dataset, but encourage submitters to apply their suggested protocols to both 
pretraining datasets.
We also suggest that submitters include model results for models trained on 
these datasets.
Suggestions are provided on the CBT website: https://genbench.org/cbt.
Given enough high-quality submissions, we aim to write a paper with the 
combined results, to which submitters can be co-authors, if they wish so.
More detailed guidelines will be given on https://genbench.org/cbt.

Archival vs extended abstract
Archival papers are up to 8 pages excluding references and report on completed, 
original and unpublished research.
They follow the requirements of regular EMNLP 2024 submissions.
Accepted papers will be published in the workshop proceedings and are expected 
to be presented at the workshop.
The papers will undergo double-blind peer review and should thus be anonymised.
Extended abstracts can be up to 2 pages excluding references, and may report on 
work in progress or be cross-submissions of work that has already appeared in 
another venue.
Abstract titles will be posted on the workshop website, but will not be 
included in the proceedings.

Submission instructions
For both archival papers and extended abstracts, we refer to the EMNLP 2024 
website for paper templates and requirements.
Additional requirements for both regular workshop papers and collaborative 
benchmarking task submissions can be found on our website.
All submissions can be submitted through OpenReview: 
https://openreview.net/group?id=GenBench.org/2024/Workshop.

Important dates
        •       August 15, 2024: Paper submission deadline
        •       September 20, 2024: Notification deadline
        •       October 4, 2024: Camera-ready deadline
        •       November 15 or 16, 2024: Workshop
Note: all deadlines are 11:59 PM UTC-12:00. Check the website for final updates 
to these deadlines (https://genbench.org/workshop).

Preprints
We do not have an anonymity deadline, preprints are allowed, both before the 
submission deadline as well as after.

Contact
Email address: [email protected]
Website: https://genbench.org/workshop

On behalf of the organisers
Dieuwke Hupkes
Verna Dankers
Khuyagbaatar Batsuren
Amirhossein Kazemnejad
Christos Christodoulopoulos
Mario Giulianelli
Ryan Cotterell
_______________________________________________
Corpora mailing list -- [email protected]
https://list.elra.info/mailman3/postorius/lists/corpora.list.elra.info/
To unsubscribe send an email to [email protected]
_______________________________________________
Corpora mailing list -- [email protected]
https://list.elra.info/mailman3/postorius/lists/corpora.list.elra.info/
To unsubscribe send an email to [email protected]

Reply via email to