DMR 2023, the Fourth International Workshop on Designing Meaning
Representations, will be co-located with IWCS 2023
<http://iwcs2023.loria.fr/>. It will be held on June 20th, 2023 in Nancy,
France.

While deep learning methods have led to many breakthroughs in practical
natural language applications, most notably in Machine Translation, Machine
Reading, Question Answering, Recognizing Textual Entailment, and so on,
there is still a sense among many NLP researchers that we have a long way
to go before we can develop systems that can actually “understand” human
language and explain the decisions they make. Indeed, “understanding”
natural language entails many different human-like capabilities, and they
include but are not limited to the ability to track entities in a text,
understand the relations between these entities, track events and their
participants described in a text, understand how events unfold in time, and
distinguish events that have actually happened from events that are planned
or intended, are uncertain, or did not happen at all. We believe a critical
step in achieving natural language understanding is to design meaning
representations for text that have the necessary meaning “ingredients” that
help us achieve these capabilities.  Such meaning representations can also
potentially be used to evaluate the compositional generalization capacity
of deep learning models.

There has been a growing body of research devoted to the design,
annotation, and parsing of meaning representations in recent years. The
meaning representations that have been used for semantic parsing research
are developed with different linguistic perspectives and practical goals in
mind and have different formal properties. Formal meaning representation
frameworks such as Minimal Recursion Semantics (MRS) and Discourse
Representation Theory (as exemplified in the Parallel Meaning Bank) are
developed with the goal of supporting logical inference in reasoning-based
AI systems and are therefore easily translatable into first-order logic,
requiring proper representation of semantic components such as
quantification, negation, tense, and modality. Other meaning representation
frameworks such as Abstract Meaning Representation (AMR), Tecto-grammatical
Representation (TR) in Prague Dependency Treebanks and the Universal
Conceptual Cognitive Annotation (UCCA), put more emphasis on the
representation of core predicate-argument structure, lexical semantic
information such as semantic roles and word senses, or named entities and
relations. There is also a more recent effort in developing a Uniform
Meaning Representation (UMR) that is based on AMR but extends it to
cross-linguistic settings and enhances it to represent document-level
semantic content. The automatic parsing of natural language text into these
meaning representations and the generation of natural language text from
these meaning representations are also very active areas of research, and a
wide range of technical approaches and learning methods have been applied
to these problems.

This workshop will bring together researchers who are producers and
consumers of meaning representations, and through their interaction develop
a deeper understanding of the key elements of meaning representations that
are the most valuable to the NLP community. The workshop will also provide
an opportunity for meaning representation researchers to critically examine
existing frameworks with the goal of using their findings to inform the
design of next-generation meaning representations. A third goal of the
workshop is to explore opportunities and identify challenges in the design
and use of meaning representations in multilingual settings. A final goal
of the workshop is to understand the relationship between distributed
meaning representations trained on large data sets using network models,
and the symbolic meaning representations that are carefully designed and
annotated by NLP researchers and gain a deeper understanding of areas where
each type of meaning representation is the most effective.

The workshop will solicit papers that address one or more of the following
topics:

— Design and annotation of meaning representations;

— Cross-framework comparison of meaning representations;

— Challenges and techniques in automatic parsing of meaning representations;

— Challenges and techniques in automatically generating text from meaning
representations;

— Meaning representation evaluation metrics;

— Lexical resources, ontologies, and grounding in relation to meaning
representations;

— Real-world applications of meaning representations;

— Issues in applying meaning representations to multilingual settings and
lower-resourced languages;

— The relationship between symbolic meaning representations and distributed
semantic representations;

— Formal properties of meaning representations;

— Any other topics that address the design, processing, and use of meaning
representations.

Important dates:

All deadlines are 11:59PM UTC-12:00 (“anywhere on Earth”).

Paper due: April 3, 2023

Notification of acceptance: May 1, 2023

Camera-ready deadline: June 1, 2023

Workshop date: June 20, 2023

Submission instructions:


Submission site: https://softconf.com/iwcs2023/dmr2023

Submissions should report original and unpublished research on topics of
interest to the workshop. Accepted papers are expected to be presented at
the workshop and will be published in the workshop proceedings on the ACL
Anthology. They should emphasize obtained results rather than intended
work, and should indicate clearly the state of completion of the reported
results. A paper accepted for presentation at the workshop must not be or
have been presented at any other meeting with publicly available
proceedings.

Submission is electronic, using the Softconf START conference management
system. Here is the link <https://softconf.com/iwcs2023/dmr2023> to the DMR
submission site. Submissions must adhere to the two-column format of ACL
venues: please use our specific style-files
<https://iwcs2021.github.io/download/iwcs2021-templates.zip> or the Overleaf
template
<https://www.overleaf.com/latex/templates/instructions-for-iwcs-2021-proceedings/fpnsyxqqpfbw>,
taken from ACL 2021. Similar to ACL 2021, initial submissions should be
fully anonymous to ensure double-blind reviewing. Long papers must not
exceed eight (8) pages of content. Short papers and demonstration papers
must not exceed four (4) pages of content. If a paper is accepted, it will
be given an additional page to address reviewers’ comments in the final
version of the paper. References and appendices do not count against these
limits.

Reviewing of papers will be double-blind. Therefore, the paper must not
include the authors' names and affiliations or self-references that reveal
the author's identity--e.g., "We previously showed (Smith, 1991) ..."
should be replaced with citations such as "Smith (1991) previously showed
...". Papers that do not conform to these requirements will be rejected
without review.

Authors of papers that have been or will be submitted to other meetings or
publications must provide this information to the workshop organizers (
[email protected]). Authors of accepted papers must notify
the program chairs within 10 days of acceptance if the paper is withdrawn
for any reason.

DMR 2023 does not have an anonymity period. However, we ask you to be
reasonable and not publicly advertise your preprint during (or right
before) review.

More information will be posted to the DMR 2023 website:
http://iwcs2023.loria.fr/dmr-2023-the-fourth-international-workshop-on-designing-meaning-representation/
_______________________________________________
Corpora mailing list -- [email protected]
https://list.elra.info/mailman3/postorius/lists/corpora.list.elra.info/
To unsubscribe send an email to [email protected]

Reply via email to