Project Details
Projekt Print View

Core Technologies for Statistical Machine Translation

Subject Area Image and Language Processing, Computer Graphics and Visualisation, Human Computer Interaction, Ubiquitous and Wearable Computing
General and Comparative Linguistics, Experimental Linguistics, Typology, Non-European Languages
Term from 2017 to 2020
Project identifier Deutsche Forschungsgemeinschaft (DFG) - Project number 327471424
 
Final Report Year 2020

Final Report Abstract

This work deals with the investigation of various core technologies for machine translation. In the general environment where neural machine translation predominates, we explore various potential network architectures for the task, rather than using the state-of-the-art approaches and making minor changes. Our novel neuralnetwork-based direct hidden Markov model with a better interpretable architecture achieves a comparable performance to the strong LSTM-RNN-based attention model. Our two-dimensional sequence-to-sequence model, a network that uses a two-dimensional LSTM unit to read both the source and the target sentence jointly, even surpasses the state-of-the-art attention baseline. Despite the drawbacks of training and decoding speed, which may affect their use in industrial applications, we still believe that these two alternative architectures give us insights into the system rather than using the neural machine translation system as a black box. On top of the transformer-based architecture, we improve word alignment quality for the purpose of dictionary-guided translation and analyze different types of positional encoding. We also propose alternative approaches of unsupervised machine translation and transfer learning that bring significant improvements over the state-of-the-art baselines. Moreover, we develop strategies for weight initialization and normalization and apply kernel functions in the softmax layer to improve the neural language models. In addition to the research results, we continuously participate in the global evaluation campaigns to test our open source tools such as Sisyphus, uniblock and Extended Edit Distance. The performance of our systems is always among the best in various types of evaluation tasks. Our punctuation prediction models perform well and play an important role in the cascaded speech-to-text translation system. In addition, the end-to-end speech translation model is now available in our speech recognition and machine translation toolkit RETURNN, with which we participate in the IWSLT 2020 speech translation task.

Publications

 
 

Additional Information

Textvergrößerung und Kontrastanpassung