Abstract
In this paper, we propose a new method for calculating the output layer in neural machine translation systems. The method is based on predicting a binary code for each word and can reduce computation time/memory requirements of the output layer to be logarithmic in vocabulary size in the best case. In addition, we also introduce two advanced approaches to improve the robustness of the proposed model: using error-correcting codes and combining softmax and binary codes. Experiments on two English ? Japanese bidirectional translation tasks show proposed models achieve BLEU scores that approach the softmax, while reducing memory usage to the order of less than 1/10 and improving decoding speed on CPUs by x5 to x10.
Original language | English |
---|---|
Title of host publication | ACL 2017 - The 55th Annual Meeting of the Association for Computational Linguistics |
Subtitle of host publication | Proceedings of the Conference, Vol. 1 (Long Papers) July 30 |
Editors | Regina Barzilay, Min-Yen Kan |
Place of Publication | Stroudsburg PA USA |
Publisher | Association for Computational Linguistics (ACL) |
Pages | 850-860 |
Number of pages | 11 |
Volume | 1 |
ISBN (Electronic) | 9781945626753 |
DOIs | |
Publication status | Published - 2017 |
Externally published | Yes |
Event | Annual Meeting of the Association of Computational Linguistics 2017 - Vancouver, Canada Duration: 30 Jul 2017 → 4 Aug 2017 Conference number: 55th https://www.aclweb.org/anthology/events/acl-2017/ (Proceedings) |
Conference
Conference | Annual Meeting of the Association of Computational Linguistics 2017 |
---|---|
Abbreviated title | ACL 2017 |
Country/Territory | Canada |
City | Vancouver |
Period | 30/07/17 → 4/08/17 |
Internet address |
|