Towards Two-dimensional Sequence To Sequence Model In Neural Machine Translation | Awesome LLM Papers

Towards Two-dimensional Sequence To Sequence Model In Neural Machine Translation

Parnia Bahar, Christopher Brix, Hermann Ney Β· Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing Β· 2018

This work investigates an alternative model for neural machine translation (NMT) and proposes a novel architecture, where we employ a multi-dimensional long short-term memory (MDLSTM) for translation modeling. In the state-of-the-art methods, source and target sentences are treated as one-dimensional sequences over time, while we view translation as a two-dimensional (2D) mapping using an MDLSTM layer to define the correspondence between source and target words. We extend beyond the current sequence to sequence backbone NMT models to a 2D structure in which the source and target sentences are aligned with each other in a 2D grid. Our proposed topology shows consistent improvements over attention-based sequence to sequence model on two WMT 2017 tasks, German(\leftrightarrow)English.

Similar Work
Loading…