ログイン 新規登録
言語:

WEKO3

  • トップ
  • ランキング
To
lat lon distance
To

Field does not validate



インデックスリンク

インデックスツリー

メールアドレスを入力してください。

WEKO

One fine body…

WEKO

One fine body…

アイテム

  1. 論文誌(ジャーナル)
  2. Vol.59
  3. No.5

Exploiting Multilingual Corpora Simply and Efficiently in Neural Machine Translation

https://ipsj.ixsq.nii.ac.jp/records/189394
https://ipsj.ixsq.nii.ac.jp/records/189394
13594f0b-f10a-4e32-925d-e597e3fb6abb
名前 / ファイル ライセンス アクション
IPSJ-JNL5905011.pdf IPSJ-JNL5905011.pdf (503.9 kB)
Copyright (c) 2018 by the Information Processing Society of Japan
オープンアクセス
Item type Journal(1)
公開日 2018-05-15
タイトル
タイトル Exploiting Multilingual Corpora Simply and Efficiently in Neural Machine Translation
タイトル
言語 en
タイトル Exploiting Multilingual Corpora Simply and Efficiently in Neural Machine Translation
言語
言語 eng
キーワード
主題Scheme Other
主題 [一般論文] Neural Machine Translation (NMT), multi-source NMT, empirical comparison, transfer learning, deep learning, dictionary extraction
資源タイプ
資源タイプ識別子 http://purl.org/coar/resource_type/c_6501
資源タイプ journal article
著者所属
Graduate School of Informatics, Kyoto University
著者所属
Japan Science and Technology Agency
著者所属
Graduate School of Informatics, Kyoto University
著者所属(英)
en
Graduate School of Informatics, Kyoto University
著者所属(英)
en
Japan Science and Technology Agency
著者所属(英)
en
Graduate School of Informatics, Kyoto University
著者名 Raj, Dabre

× Raj, Dabre

Raj, Dabre

Search repository
Fabien, Cromieres

× Fabien, Cromieres

Fabien, Cromieres

Search repository
Sadao, Kurohashi

× Sadao, Kurohashi

Sadao, Kurohashi

Search repository
著者名(英) Raj, Dabre

× Raj, Dabre

en Raj, Dabre

Search repository
Fabien, Cromieres

× Fabien, Cromieres

en Fabien, Cromieres

Search repository
Sadao, Kurohashi

× Sadao, Kurohashi

en Sadao, Kurohashi

Search repository
論文抄録
内容記述タイプ Other
内容記述 In this paper, we explore a simple approach for “Multi-Source Neural Machine Translation” (MSNMT) which only relies on preprocessing a N-way multilingual corpus without modifying the Neural Machine Translation (NMT) architecture or training procedure. We simply concatenate the source sentences to form a single, long multi-source input sentence while keeping the target side sentence as it is and train an NMT system using this preprocessed corpus. We evaluate our method in resource poor as well as resource rich settings and show its effectiveness (up to 4 BLEU using 2 source languages and up to 6 BLEU using 5 source languages) and compare them against existing approaches. We also provide some insights on how the NMT system leverages multilingual information in such a scenario by visualizing attention. We then show that this multi-source approach can be used for transfer learning to improve the translation quality for single-source systems without using any additional corpora thereby highlighting the importance of multilingual-multiway corpora in low resource scenarios. We also extract and evaluate a multilingual dictionary by a method that utilizes the multi-source attention and show that it works fairly well despite its simplicity.
------------------------------
This is a preprint of an article intended for publication Journal of
Information Processing(JIP). This preprint should not be cited. This
article should be cited as: Journal of Information Processing Vol.26(2018) (online)
DOI http://dx.doi.org/10.2197/ipsjjip.26.406
------------------------------
論文抄録(英)
内容記述タイプ Other
内容記述 In this paper, we explore a simple approach for “Multi-Source Neural Machine Translation” (MSNMT) which only relies on preprocessing a N-way multilingual corpus without modifying the Neural Machine Translation (NMT) architecture or training procedure. We simply concatenate the source sentences to form a single, long multi-source input sentence while keeping the target side sentence as it is and train an NMT system using this preprocessed corpus. We evaluate our method in resource poor as well as resource rich settings and show its effectiveness (up to 4 BLEU using 2 source languages and up to 6 BLEU using 5 source languages) and compare them against existing approaches. We also provide some insights on how the NMT system leverages multilingual information in such a scenario by visualizing attention. We then show that this multi-source approach can be used for transfer learning to improve the translation quality for single-source systems without using any additional corpora thereby highlighting the importance of multilingual-multiway corpora in low resource scenarios. We also extract and evaluate a multilingual dictionary by a method that utilizes the multi-source attention and show that it works fairly well despite its simplicity.
------------------------------
This is a preprint of an article intended for publication Journal of
Information Processing(JIP). This preprint should not be cited. This
article should be cited as: Journal of Information Processing Vol.26(2018) (online)
DOI http://dx.doi.org/10.2197/ipsjjip.26.406
------------------------------
書誌レコードID
収録物識別子タイプ NCID
収録物識別子 AN00116647
書誌情報 情報処理学会論文誌

巻 59, 号 5, 発行日 2018-05-15
ISSN
収録物識別子タイプ ISSN
収録物識別子 1882-7764
戻る
0
views
See details
Views

Versions

Ver.1 2025-01-20 01:37:58.006455
Show All versions

Share

Mendeley Twitter Facebook Print Addthis

Cite as

エクスポート

OAI-PMH
  • OAI-PMH JPCOAR
  • OAI-PMH DublinCore
  • OAI-PMH DDI
Other Formats
  • JSON
  • BIBTEX

Confirm


Powered by WEKO3


Powered by WEKO3