Domain generalisation of NMT: fusing adapters with leave-one-domain-out training

Research output: Chapter in Book/Report/Conference proceedingConference PaperResearchpeer-review

8 Citations (Scopus)

Abstract

Generalising to unseen domains is under-explored and remains a challenge in neural machine translation. Inspired by recent research in parameter-efficient transfer learning from pretrained models, this paper proposes a fusion-based generalisation method that learns to combine domain-specific parameters. We propose a leave-one-domain-out training strategy to avoid information leaking to address the challenge of not knowing the test domain during training time. Empirical results on three language pairs show that our proposed fusion method outperforms other baselines up to +0.8 BLEU score on average.

Original languageEnglish
Title of host publicationACL 2022 - The 60th Annual Meeting of the Association for Computational Linguistics - Findings of ACL 2022
EditorsSmaranda Muresan, Preslav Nakov, Aline Villavicencio
Place of PublicationStroudsburg PA USA
PublisherAssociation for Computational Linguistics (ACL)
Pages582-588
Number of pages7
ISBN (Electronic)9781955917254
DOIs
Publication statusPublished - 2022
EventAnnual Meeting of the Association of Computational Linguistics 2022 - Dublin, Ireland
Duration: 22 May 202227 May 2022
Conference number: 60th
https://aclanthology.org/volumes/2022.acl-short/ (Proceedings - Short)
https://aclanthology.org/volumes/2022.acl-long/ (Proceedings - Long)
https://www.2022.aclweb.org/ (Website)

Publication series

NameProceedings of the Annual Meeting of the Association for Computational Linguistics
PublisherAssociation for Computational Linguistics (ACL)
ISSN (Print)0736-587X

Conference

ConferenceAnnual Meeting of the Association of Computational Linguistics 2022
Abbreviated titleACL 2022
Country/TerritoryIreland
CityDublin
Period22/05/2227/05/22
Internet address

Cite this