Incorporating Source-Side Phrase Structures into Neural Machine Translation

Authors

  • Akiko Eriguchi Microsoft Research
  • Kazuma Hashimoto Salesforce Research
  • Yoshimasa Tsuruoka The University of Tokyo

Abstract

Neural Machine Translation (NMT) has shown great success as a fully Statistical Machine Translation (SMT) model in several languages. Early NMT models are based on sequence-to-sequence learning that encodes a sequence of source words into a vector space and generates another sequence of target words from the vector. In those NMT models, sentences are simply treated as sequences of words without any internal structure. In this article, we focus on the role of the syntactic structure of source sentences and propose a novel end-to-end syntactic NMT model, which we call a tree-to-sequence NMT model, extending a sequence-to-sequence model with the source-side phrase structure. Our proposed model has an attention mechanism that enables the decoder to generate a translated word while softly aligning it with phrases as well as words of the source sentence. Experimenting on Chinese-to-Japanese and English-to-Japanese translation tasks, we report different trends between the sequence-to-sequence and tree-to-sequence models. Experimental results on the English-to-Japanese dataset demonstrate that our proposed model considerably outperforms the sequence-to-sequence NMT models and compares favorably with the state-of-the-art tree-to-string SMT system.

Author Biographies

  • Akiko Eriguchi, Microsoft Research

    Akiko Eriguchi is a senior research scientist at Microsoft Research. She received her B.S. and M.S. degrees from Ochanomizu University in 2013 and 2015, respectively, and her Ph.D. degrees from the University of Tokyo in 2018. Her major research interests include natural
    language processing and machine learning.

  • Kazuma Hashimoto, Salesforce Research

    Kazuma Hashimoto is a research scientist at Salesforce Research. He received his B.S., M.S., and Ph.D. degrees from the University of Tokyo in 2013, 2015, and 2018, respectively. His major research interests are in natural language processing with neural networks, especially for syntax-based models and multi-task learning.

  • Yoshimasa Tsuruoka, The University of Tokyo

    Yoshimasa Tsuruoka is a professor at Department of Information and Communication Engineering, the University of Tokyo. He received his B.E., M.E. and Ph.D. degrees from University of Tokyo in 1997, 1999 and 2002. His major research interests include natural language processing, text mining from biomedical literature and artificial intelligence in games.

Published

2024-12-05

Issue

Section

Short paper