Abstract
Out-of-vocabulary words present a great challenge for Machine Translation. Recently various character-level compositional models were proposed to address this issue. In current research we incorporate two most popular neural architectures, namely LSTM and CNN, into hard- and soft-attentional models of translation for character-level representation of the source. We propose semantic and morphological intrinsic evaluation of encoder-level representations. Our analysis of the learned representations reveals that character-based LSTM seems to be better at capturing morphological aspects compared to character-based CNN. We also show that a hard-attentional model provides better character-level representations compared to standard 'soft' attention.
| Original language | English |
|---|---|
| Title of host publication | EMNLP 2017 - First Workshop on Subword and Character Level Models in NLP, Proceedings of the Workshop |
| Editors | Manaal Faruqui, Hinrich Schutze, Isabel Trancoso, Yaghoobzadeh Yadollah |
| Place of Publication | Stroudsburg PA USA |
| Publisher | Association for Computational Linguistics (ACL) |
| Pages | 103-108 |
| Number of pages | 6 |
| ISBN (Electronic) | 9781945626913 |
| DOIs | |
| Publication status | Published - 2017 |
| Event | Workshop on Subword and Character Level Models in NLP 2017 - Copenhagen, Denmark Duration: 7 Sept 2017 → 7 Sept 2017 Conference number: 1st https://aclanthology.org/volumes/W17-41/ (Proceedings) |
Workshop
| Workshop | Workshop on Subword and Character Level Models in NLP 2017 |
|---|---|
| Abbreviated title | SCLeM 2017 |
| Country/Territory | Denmark |
| City | Copenhagen |
| Period | 7/09/17 → 7/09/17 |
| Internet address |
|
Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver