Abstract
Out-of-vocabulary words present a great challenge for Machine Translation. Recently various character-level compositional models were proposed to address this issue. In current research we incorporate two most popular neural architectures, namely LSTM and CNN, into hard- and soft-attentional models of translation for character-level representation of the source. We propose semantic and morphological intrinsic evaluation of encoder-level representations. Our analysis of the learned representations reveals that character-based LSTM seems to be better at capturing morphological aspects compared to character-based CNN. We also show that a hard-attentional model provides better character-level representations compared to standard 'soft' attention.
Original language | English |
---|---|
Title of host publication | EMNLP 2017 - First Workshop on Subword and Character Level Models in NLP, Proceedings of the Workshop |
Editors | Manaal Faruqui, Hinrich Schutze, Isabel Trancoso, Yaghoobzadeh Yadollah |
Place of Publication | Stroudsburg PA USA |
Publisher | Association for Computational Linguistics (ACL) |
Pages | 103-108 |
Number of pages | 6 |
ISBN (Electronic) | 9781945626913 |
DOIs | |
Publication status | Published - 2017 |
Event | Workshop on Subword and Character Level Models in NLP 2017 - Copenhagen, Denmark Duration: 7 Sep 2017 → 7 Sep 2017 Conference number: 1st https://aclanthology.org/volumes/W17-41/ (Proceedings) |
Workshop
Workshop | Workshop on Subword and Character Level Models in NLP 2017 |
---|---|
Abbreviated title | SCLeM 2017 |
Country/Territory | Denmark |
City | Copenhagen |
Period | 7/09/17 → 7/09/17 |
Internet address |
|