Bibert Ende
BiBERT-ende is a bilingual (English-German) pretrained language model optimized for Neural Machine Translation (NMT), enhancing translation performance through contextual embeddings.
Machine Translation
Transformers Supports Multiple Languages#Bilingual Pretraining#Neural Machine Translation#Contextual Embeddings

Downloads 40
Release Time : 3/2/2022
Model Overview
BiBERT-ende is a customized bilingual pretrained language model designed to simplify the integration of existing pretrained models by directly feeding contextual embeddings into the NMT encoder, achieving state-of-the-art translation performance.
Model Features
Bilingual Pretraining
Specifically pretrained for English and German, optimizing cross-lingual contextual understanding.
Simplified Integration
Simplifies the integration of pretrained models by directly using contextual embeddings as NMT encoder input.
Random Layer Selection
Introduces random layer selection to ensure full utilization of different levels of contextual embeddings.
Bidirectional Translation Model
Supports bidirectional translation (EnglishāGerman and GermanāEnglish) with high performance in both directions.
Model Capabilities
English-to-German machine translation
German-to-English machine translation
Contextual embedding generation
Use Cases
Machine Translation
IWSLT'14 Dataset Translation
Achieves BLEU scores of 30.45 for EnglishāGerman and 38.61 for GermanāEnglish on the IWSLT'14 dataset.
Surpasses all published results
WMT'14 Dataset Translation
Achieves BLEU scores of 31.26 for EnglishāGerman and 34.94 for GermanāEnglish on the WMT'14 dataset.
Surpasses all published results
Featured Recommended AI Models
Ā© 2025AIbase