B

Bibert Ende

Developed by jhu-clsp
BiBERT-ende is a bilingual (English-German) pretrained language model optimized for Neural Machine Translation (NMT), enhancing translation performance through contextual embeddings.
Downloads 40
Release Time : 3/2/2022

Model Overview

BiBERT-ende is a customized bilingual pretrained language model designed to simplify the integration of existing pretrained models by directly feeding contextual embeddings into the NMT encoder, achieving state-of-the-art translation performance.

Model Features

Bilingual Pretraining
Specifically pretrained for English and German, optimizing cross-lingual contextual understanding.
Simplified Integration
Simplifies the integration of pretrained models by directly using contextual embeddings as NMT encoder input.
Random Layer Selection
Introduces random layer selection to ensure full utilization of different levels of contextual embeddings.
Bidirectional Translation Model
Supports bidirectional translation (English→German and German→English) with high performance in both directions.

Model Capabilities

English-to-German machine translation
German-to-English machine translation
Contextual embedding generation

Use Cases

Machine Translation
IWSLT'14 Dataset Translation
Achieves BLEU scores of 30.45 for English→German and 38.61 for German→English on the IWSLT'14 dataset.
Surpasses all published results
WMT'14 Dataset Translation
Achieves BLEU scores of 31.26 for English→German and 34.94 for German→English on the WMT'14 dataset.
Surpasses all published results
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Ā© 2025AIbase