Joint Training for Neural Machine Translation

This book presents four approaches to jointly training bidirectional neural machine translation (NMT) models. First, in order to improve the accuracy of the attention mechanism, it proposes an agreement-based joint training approach to help the two complementary models agree on word alignment matric...

Full description

Main Author: Cheng, Yong. (Author, http://id.loc.gov/vocabulary/relators/aut)
Corporate Author: SpringerLink (Online service)
Language:English
Published: Singapore : Springer Singapore : Imprint: Springer, 2019.
Edition:1st ed. 2019.
Series:Springer Theses, Recognizing Outstanding Ph.D. Research,
Subjects:
Online Access:https://doi.org/10.1007/978-981-32-9748-7
Summary:This book presents four approaches to jointly training bidirectional neural machine translation (NMT) models. First, in order to improve the accuracy of the attention mechanism, it proposes an agreement-based joint training approach to help the two complementary models agree on word alignment matrices for the same training data. Second, it presents a semi-supervised approach that uses an autoencoder to reconstruct monolingual corpora, so as to incorporate these corpora into neural machine translation. It then introduces a joint training algorithm for pivot-based neural machine translation, which can be used to mitigate the data scarcity problem. Lastly it describes an end-to-end bidirectional NMT model to connect the source-to-target and target-to-source translation models, allowing the interaction of parameters between these two directional models.
Physical Description:XIII, 78 p. 23 illus., 9 illus. in color. online resource.
ISBN:9789813297487
ISSN:2190-5053