Multi-Source Simultaneous Speech Translation
Thesis title in Czech: | Simultánní překlad řeči z více zdrojů |
---|---|
Thesis title in English: | Multi-Source Simultaneous Speech Translation |
Key words: | simultánní překlad řeči|překlad řeči|překlad z řeči do textu|strojový překlad|vícejazyčnost|vícezdrojovost|simultánní tlumočení|zpracování přirozeného jazyka |
English key words: | simultaneous speech translation|speech translation|speech-to-text translation|machine translation|multilinguality|multi-sourcing|simultaneous interpreting|natural language processing |
Academic year of topic announcement: | 2018/2019 |
Thesis type: | dissertation |
Thesis language: | angličtina |
Department: | Institute of Formal and Applied Linguistics (32-UFAL) |
Supervisor: | doc. RNDr. Ondřej Bojar, Ph.D. |
Author: | hidden - assigned and confirmed by the Study Dept. |
Date of registration: | 17.06.2019 |
Date of assignment: | 17.06.2019 |
Confirmed by Study dept. on: | 04.10.2019 |
Date and time of defence: | 06.06.2024 10:00 |
Date of electronic submission: | 11.03.2024 |
Date of submission of printed version: | 22.03.2024 |
Guidelines |
Neural machine translation (NMT) has the capability of handling more source and/or target languages at once.
The goal of the thesis is to experimentally explore this area and propose and evaluate variations of NMT model architectures, training data layout or training methods to achieve gains in translation quality or efficiency. Depending on the results of experiments carried out in the first stage of the studies, the work may focus primarily on one of the following use cases: - Multi-target MT, where the same input is to be translated simultaneously into multiple languages. The desired savings would be primarily in terms of memory and computing resources (one model serving more target languages), at as little loss in translation quality as possible. Thanks to GPU parallelization, multiple target languages could be produced synchronously. - Multi-source MT, where the same input sentence is available in more than one language. The expected gains would be in translation quality, thanks to the reduced ambiguity of the input. The proposed architecture should gracefully handle if not all input language versions are available, including the situation where only one source language is given. |
References |
Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is All you Need. In I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, editors, Advances in Neural Information Processing Systems 30, pages 6000–6010. Curran Associates, Inc., 2017.
Orhan Firat, Kyunghyun Cho, and Yoshua Bengio. Multi-way, multilingual neural machine translation with a shared attention mechanism. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 866–875, San Diego, California, June 2016. Association for Computational Linguistics. Goodfellow, I., Y. Bengio, and A. Courville 2016. Deep learning. Cambridge, MA, USA: MIT press. Helcl Jindřich, Libovický Jindřich, Kocmi Tom, Musil Tomáš, Cífka Ondřej, Variš Dušan, Bojar Ondřej: Neural Monkey: The Current State and Beyond. In: The 13th Conference of The Association for Machine Translation in the Americas, Vol. 1: MT Researchers’ Track, Copyright © The Association for Machine Translation in the Americas, Stroudsburg, PA, USA, pp. 168-176, 2018 |