Hi all,

This is just a reminder that the expression of interest for this
volume is due in less than a week!

The expression of interest is easy: just a title, list of authors, and
a short description.

If anyone else would like to help out with the updated Apertium paper
that we're planning to submit, then please get in touch.

--
Jonathan

пт, 1 нояб. 2019 г. в 22:21, Jonathan Washington
<jonathan.n.washing...@gmail.com>:
>
> Hi all,
>
> Below please find a revised CFP for the Machine Translation Special
> Issue on MT for Low-Resource Languages.
>
> =====
> CALL FOR PAPERS: Machine Translation Journal
> Special Issue on Machine Translation for Low-Resource Languages
> https://www.springer.com/computer/ai/journal/10590/
>
> GUEST EDITORS (Listed alphabetically)
> • Alina Karakanta (FBK-Fondazione Bruno Kessler)
> • Audrey N. Tong (NIST)
> • Chao-Hong Liu (ADAPT Centre/Dublin City University)
> • Ian Soboroff (NIST)
> • Jonathan Washington (Swarthmore College)
> • Oleg Aulov (NIST)
> • Xiaobing Zhao (Minzu University of China)
>
> Machine translation (MT) technologies have been improved significantly
> in the last two decades, with developments in phrase-based statistical
> MT (SMT) and recently neural MT (NMT). However, most of these methods
> rely on the availability of large parallel data for training the MT
> systems, resources which are not available for the majority of
> language pairs, and hence current technologies often fall short in
> their ability to be applied to low-resource languages. Developing MT
> technologies using relatively small corpora still presents a major
> challenge for the MT community. In addition, many methods for
> developing MT systems still rely on several natural language
> processing (NLP) tools to pre-process texts in source languages and
> post-process MT outputs in target languages. The performance of these
> tools often has a great impact on the quality of the resulting
> translation. The availability of MT technologies and NLP tools can
> facilitate equal access to information for the speakers of a language
> and determine on which side of the digital divide they will end up.
> The lack of these technologies for many of the world's languages
> provides opportunities both for the field to grow and for making tools
> available for speakers of low-resource languages.
>
> In recent years, several workshops and evaluations have been organized
> to promote research on low-resource languages. NIST has been
> conducting Low Resource Human Language Technology evaluations
> (LoReHLT) annually from 2016 to 2019. In LoReHLT evaluations, there is
> no training data in the evaluation language. Participants receive
> training data in related languages, but need to bootstrap systems in
> the surprise evaluation language at the start of the evaluation.
> Methods for this include pivoting approaches and taking advantage of
> linguistic universals. The evaluations are supported by DARPA's Low
> Resource Languages for Emergent Incidents (LORELEI) program, which
> seeks to advance technologies that are less dependent on large data
> resources and that can be quickly pivoted to new languages within a
> very short amount of time so that information from any language can be
> extracted in a timely manner to provide situation awareness to
> emergent incidents. There are also the Workshop on Technologies for MT
> of Low-Resource Languages (LoResMT) and the Workshop on Deep Learning
> Approaches for Low-Resource Natural Language Processing (DeepLo),
> which provide a venue for sharing research and working on the research
> and development in this field.
>
> This special issue solicits original research papers on MT
> systems/methods and related NLP tools for low-resource languages in
> general. LoReHLT, LORELEI, LoResMT and DeepLo participants are very
> welcome to submit their work to the special issue. Summary papers on
> MT research for specific low-resource languages, as well as extended
> versions (>40% difference) of published papers from relevant
> conferences/workshops are also welcome.
>
> Topics of the special issue include but are not limited to:
>  * Research and review papers of MT systems/methods for low-resource languages
>  * Research and review papers of pre-processing and/or post-processing
> NLP tools for MT
>  * Word tokenizers/de-tokenizers for low-resource languages
>  * Word/morpheme segmenters for low-resource languages
>  * Use of morphological analyzers and/or morpheme segmenters in MT
>  * Multilingual/cross-lingual NLP tools for MT
>  * Review of available corpora of low-resource languages for MT
>  * Pivot MT for low-resource languages
>  * Zero-shot MT for low-resource languages
>  * Fast building of MT systems for low-resource languages
>  * Re-usability of existing MT systems and/or NLP tools for
> low-resource languages
>  * Machine translation for language preservation
>  * Techniques that work across many languages and modalities
>  * Techniques that are less dependent on large data resources
>  * Use of language-universal resources
>  * Bootstrap trained resources for short development cycle
>  * Entity-, relation- and event-extraction
>  * Sentiment detection
>  * Summarization
>  * Processing diverse languages, genres (news, social media, etc.) and
> modalities (text, speech, video, etc.)
>
> IMPORTANT DATES
> November 26, 2019: Expression of interest (EOI)
> February 25, 2020: Paper submission deadline
> July 7, 2020: Camera-ready papers due
> December, 2020: Publication
>
> SUBMISSION GUIDELINES
> o For EOI, please submit via the link: https://forms.gle/mAQH4qaPTuzDhEceA
> o For paper submission, please go to the MT journal website
> https://link.springer.com/journal/10590 and select this special issue
> o Authors should follow the "Instructions for Authors"
> o Recommended length of paper is 15 pages
> =====
>
> --
> Jonathan


_______________________________________________
Apertium-stuff mailing list
Apertium-stuff@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/apertium-stuff

Reply via email to