Thesis (Selection of subject)Thesis (Selection of subject)(version: 368)
Thesis details
   Login via CAS
Multi-Source Simultaneous Speech Translation
Thesis title in Czech: Simultánní překlad řeči z více zdrojů
Thesis title in English: Multi-Source Simultaneous Speech Translation
Key words: simultánní překlad řeči|překlad řeči|překlad z řeči do textu|strojový překlad|vícejazyčnost|vícezdrojovost|simultánní tlumočení|zpracování přirozeného jazyka
English key words: simultaneous speech translation|speech translation|speech-to-text translation|machine translation|multilinguality|multi-sourcing|simultaneous interpreting|natural language processing
Academic year of topic announcement: 2018/2019
Thesis type: dissertation
Thesis language: angličtina
Department: Institute of Formal and Applied Linguistics (32-UFAL)
Supervisor: doc. RNDr. Ondřej Bojar, Ph.D.
Author: hidden - assigned and confirmed by the Study Dept.
Date of registration: 17.06.2019
Date of assignment: 17.06.2019
Confirmed by Study dept. on: 04.10.2019
Date and time of defence: 06.06.2024 10:00
Date of electronic submission:11.03.2024
Date of submission of printed version:22.03.2024
Guidelines
Neural machine translation (NMT) has the capability of handling more source and/or target languages at once.

The goal of the thesis is to experimentally explore this area and propose and evaluate variations of NMT model architectures, training data layout or training methods to achieve gains in translation quality or efficiency. Depending on the results of experiments carried out in the first stage of the studies, the work may focus primarily on one of the following use cases:

- Multi-target MT, where the same input is to be translated simultaneously into multiple languages. The desired savings would be primarily in terms of memory and computing resources (one model serving more target languages), at as little loss in translation quality as possible. Thanks to GPU parallelization, multiple target languages could be produced synchronously.

- Multi-source MT, where the same input sentence is available in more than one language. The expected gains would be in translation quality, thanks to the reduced ambiguity of the input. The proposed architecture should gracefully handle if not all input language versions are available, including the situation where only one source language is given.
References
Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is All you Need. In I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, editors, Advances in Neural Information Processing Systems 30, pages 6000–6010. Curran Associates, Inc., 2017.

Orhan Firat, Kyunghyun Cho, and Yoshua Bengio. Multi-way, multilingual neural machine translation with a shared attention mechanism. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 866–875, San Diego, California, June 2016. Association for Computational Linguistics.

Goodfellow, I., Y. Bengio, and A. Courville 2016. Deep learning. Cambridge, MA, USA: MIT press.

Helcl Jindřich, Libovický Jindřich, Kocmi Tom, Musil Tomáš, Cífka Ondřej, Variš Dušan, Bojar Ondřej: Neural Monkey: The Current State and Beyond. In: The 13th Conference of The Association for Machine Translation in the Americas, Vol. 1: MT Researchers’ Track, Copyright © The Association for Machine Translation in the Americas, Stroudsburg, PA, USA, pp. 168-176, 2018
 
Charles University | Information system of Charles University | http://www.cuni.cz/UKEN-329.html