Detailseite
Projekt Druckansicht

Kerntechnologien für statistische maschinelle Übersetzung

Fachliche Zuordnung Bild- und Sprachverarbeitung, Computergraphik und Visualisierung, Human Computer Interaction, Ubiquitous und Wearable Computing
Allgemeine und Vergleichende Sprachwissenschaft, Experimentelle Linguistik, Typologie, Außereuropäische Sprachen
Förderung Förderung von 2017 bis 2020
Projektkennung Deutsche Forschungsgemeinschaft (DFG) - Projektnummer 327471424
 
Erstellungsjahr 2020

Zusammenfassung der Projektergebnisse

This work deals with the investigation of various core technologies for machine translation. In the general environment where neural machine translation predominates, we explore various potential network architectures for the task, rather than using the state-of-the-art approaches and making minor changes. Our novel neuralnetwork-based direct hidden Markov model with a better interpretable architecture achieves a comparable performance to the strong LSTM-RNN-based attention model. Our two-dimensional sequence-to-sequence model, a network that uses a two-dimensional LSTM unit to read both the source and the target sentence jointly, even surpasses the state-of-the-art attention baseline. Despite the drawbacks of training and decoding speed, which may affect their use in industrial applications, we still believe that these two alternative architectures give us insights into the system rather than using the neural machine translation system as a black box. On top of the transformer-based architecture, we improve word alignment quality for the purpose of dictionary-guided translation and analyze different types of positional encoding. We also propose alternative approaches of unsupervised machine translation and transfer learning that bring significant improvements over the state-of-the-art baselines. Moreover, we develop strategies for weight initialization and normalization and apply kernel functions in the softmax layer to improve the neural language models. In addition to the research results, we continuously participate in the global evaluation campaigns to test our open source tools such as Sisyphus, uniblock and Extended Edit Distance. The performance of our systems is always among the best in various types of evaluation tasks. Our punctuation prediction models perform well and play an important role in the cascaded speech-to-text translation system. In addition, the end-to-end speech translation model is now available in our speech recognition and machine translation toolkit RETURNN, with which we participate in the IWSLT 2020 speech translation task.

Projektbezogene Publikationen (Auswahl)

 
 

Zusatzinformationen

Textvergrößerung und Kontrastanpassung