The official repo for the tutorial "Linguistically Motivated Neural Machine Translation" in The 25th Annual Conference of the European Association for Machine Translation (EAMT 2024).
Part 1: Introduction slide | |
Part 2: Linguistic Features and Encoder slide | |
Part 3: Subword, Decoder, Evaluation slide |
Haiyue Song is a technical researcher at the Advanced Translation Technology Laboratory, National Institute of Information and Communications Technology (NICT), Japan. He obtained his Ph.D. at Kyoto University. His research interests include machine translation, large language models, subword segmentation, and decoding algorithms. He has MT and LLMs related publications in TALLIP, AACL, LREC, ACL, and EMNLP. | |
Hour Kaing is a researcher at the Advanced Translation Technology Laboratory, National Institute of Information and Communications Technology (NICT), Japan. He received his B.S. from Institute of Technology of Cambodia, Cambodia, his M.Sc from University of Grenoble 1, France, and his Ph.D. from NARA Institute of Science and Technology, Japan. He is interested in linguistic analysis, low-resource machine translation, language modeling, and speech processing. He has publications in TALLIP, EACL, PACLIC, LREC, and IWSLT. | |
Raj Dabre is a senior researcher at the Advanced Translation Technology Laboratory, National Institute of Information and Communications Technology (NICT), Japan and an Adjunct Faculty at IIT Madras, India. He received his Ph.D. from Kyoto University and Masters from IIT Bombay. His primary interests are in low-resource NLP, language modeling and efficiency. He has published in ACL, EMNLP, NAACL, TMLR, AAAI, AACL, IJCNLP and CSUR. |
Time | |
---|---|
9:00-9:20 | Introduction |
9:20-10:20 | Augmenting NMT Architectures with Linguistic Features |
10:20-10:50 | Coffee break |
10:50-11:20 | Linguistically Motivated Tokenization and Transfer Learning |
11:20-11:40 | Linguistically Aware Decoding |
11:40-12:00 | Linguistically Motivated Evaluation |
12:00-12:15 | Conclusions |
12:15-12:30 | QA |
The tutorial focuses on incorporating linguistics into different stages of the neural machine translation (NMT) pipeline, from pre-processing to model training to evaluation.
For machine translation (MT) tasks, purely data-driven approaches have been dominant in recent years, and language knowledge-related approaches are often neglected. This tutorial aims to highlight the importance of linguistic knowledge, especially for low-resource languages where training data is limited.
- Introduction to Neural Machine Translation
- Linguistically Motivated Tokenization and Transfer Learning
- Augmenting NMT Architectures with Linguistic Features
- Linguistically Aware Decoding
- Linguistically Motivated Evaluation
- Limitations and Future Directions
- Summary and Conclusion
- Discussion and Q/A
- Neural Machine Translation: Basics, Practical Aspects and Recent Trends - Dabre et al., 2017
- Attention is All you Need - Vaswani et al., 2017
- Neural Machine Translation by Jointly Learning to Align and Translate - Bahdanau et al., 2016
- Juman++: A Morphological Analysis Toolkit for Scriptio Continua - Tolmachev et al., 2018
- Dynamic Programming Encoding for Subword Segmentation in Neural Machine Translation - He et al., 2020
- BERTSeg: BERT Based Unsupervised Subword Segmentation for Neural Machine Translation - Song et al., 2022
- MorphyNet: a Large Multilingual Database of Derivational and Inflectional Morphology - Batsuren et al., 2021
- Linguistically Motivated Vocabulary Reduction for Neural Machine Translation from Turkish to English - Ataman et al., 2017
- Linguistically Motivated Subwords for English-Tamil Translation: University of Groningen’s Submission to WMT-2020 - Dhar et al., 2020
- Neural Machine Translation of Logographic Languages Using Sub-character Level Information
- On Romanization for Model Transfer Between Scripts in Neural Machine Translation - Amrhein and Sennrich, 2020
- RomanSetu: Efficiently unlocking multilingual capabilities of Large Language Models models via Romanization - Husain et al., 2024
- CharSpan: Utilizing Lexical Similarity to Enable Zero-Shot Machine Translation for Extremely Low-resource Languages - Maurya et al., 2024
- SelectNoise: Unsupervised Noise Injection to Enable Zero-Shot Machine Translation for Extremely Low-resource Languages - Brahma et al., 2023
- Pre-training via Leveraging Assisting Languages for Neural Machine Translation - Song et al., 2020
- IndicTrans2: Towards High-Quality and Accessible Machine Translation Models for all 22 Scheduled Indian Languages - Gala et al., 2023
- IndicBART: A Pre-trained Model for Natural Language Generation of Indic Languages - Dabre et al., 2021
- Linguistic Input Features Improve Neural Machine Translation - Sennrich and Haddow, 2016
- FeatureBART: Feature Based Sequence-to-Sequence Pre-Training for Low-Resource NMT - Chakrabarty et al., 2022
- Improving Low-Resource NMT through Relevance Based Linguistic Features Incorporation - Chakrabarty et al., 2020
- Low-Resource Multilingual Neural Translation Using Linguistic Feature Based Relevance Mechanisms - Chakrabarty et al., 2023
- Exploiting Linguistic Resources for Neural Machine Translation Using Multi-task Learning - Niehues and Cho, 2017
- Syntax-Enhanced Neural Machine Translation with Syntax-Aware Word Representations - Zhang et al., 2019
- Dependency-to-Dependency Neural Machine Translation - Wu et al., 2018
- Multi-Source Syntactic Neural Machine Translation - Currey and Heafield, 2018
- Incorporating Source Syntax into Transformer-Based Neural Machine Translation - Currey and Heafield, 2019
- Enhancing Machine Translation with Dependency-Aware Self-Attention - Bugliarello and Okazaki, 2020
- Passing Parser Uncertainty to the Transformer: Labeled Dependency Distributions for Neural Machine Translation - Pu and Sima'an, 2022
- Modeling Source Syntax for Neural Machine Translation - Li et al., 2017
- Tree-to-Sequence Attentional Neural Machine Translation - Eriguchi et al., 2016
- Improved Neural Machine Translation with a Syntax-Aware Encoder and Decoder - Chen et al., 2017
- Learning to Parse and Translate Improves Neural Machine Translation - Eriguchi et al., 2017
- Sequence-to-Dependency Neural Machine Translation - Wu et al., 2017
- A Tree-based Decoder for Neural Machine Translation - Wang et al., 2018
- Towards String-To-Tree Neural Machine Translation - Aharoni and Goldberg, 2017
- Predicting Target Language CCG Supertags Improves Neural Machine Translation - Nǎdejde et al., 2017
- Improving Neural Machine Translation with Soft Template Prediction - Yang et al., 2020
- Explicit Syntactic Guidance for Neural Text Generation - Li et al., 2023
- Linguistic Evaluation for the 2021 State-of-the-art Machine Translation Systems for German to English and English to German - Macketanz et al., 2021
- Linguistically Motivated Evaluation of Machine Translation Metrics Based on a Challenge Set - Avramidis and Macketanz, 2022
- Linguistically Motivated Evaluation of the 2023 State-of-the-art Machine Translation: Can ChatGPT Outperform NMT? - Manakhimova et al, 2023
Haiyue Song, Hour Kaing, Raj Dabre
National Institute of Information and Communications Technology (NICT)
Hikaridai 3-5, Seika-cho, Soraku-gun, Kyoto, Japan
Emails:
- Haiyue Song: [email protected]
- Hour Kaing: [email protected]
- Raj Dabre: [email protected]
@article{linguistic-mt24,
title={Linguistically Motivated Neural Machine Translation},
author={Song, Haiyue and Kaing, Hour and Dabre, Raj},
booktitle={The 25th Annual Conference of the European Association for Machine Translation (EAMT 2024)},
year={2024}
}