Code for paper Extrapolating Multilingual Language Understanding Models as Multilingual Language Generators In this paper, we propose a Semantic Guided Alignmnt-then-Denoising (SGA) approach, which adapts a pretrained multilingual understanding model, XLM-R, to a non-autoregressive language generator in a parameter-efficient way.
Unrecovering a blank sentence to target translation bt iterative refinement
- python == 3.7.11
- torch >= 1.9
- thumt == 1.2.0
- fairseq == 0.12.2
- transformers >= 4.27.1
- sentencepiece == 0.1.99
Download TED dataset via
bash data/download_ML50_v1.sh
bash preprocess_ML50_v1.sh
bash scripts/train.sh
bash scripts/inference.sh
We Update the results of mBART + prompt-tuning to show that XLM-R can achieve comparable performance with mBART in translation under the adaptation setting.
@article{wu2023extrapolating,
title={Extrapolating Multilingual Understanding Models as Multilingual Generators},
author={Wu, Bohong and
Yuan, Fei and
Zhao, Hai and
Li, Lei and
Xu, Jingjing},
journal={findings of EMNLP 2023},
year={2023}
}