G

Gte En Mlm Large

Developed by Alibaba-NLP
Large English text encoder in the GTE-v1.5 series, supporting context lengths up to 8192, built on an improved BERT architecture.
Downloads 171
Release Time : 8/6/2024

Model Overview

This model is a general-purpose text encoder developed by Alibaba Group's Institute of Intelligent Computing, primarily used for English text embedding representation and re-ranking tasks, supporting long-context processing.

Model Features

Long-context support
Supports context lengths up to 8192, suitable for processing long documents and complex texts.
Improved BERT architecture
Enhanced architecture combining RoPE and GLU, improving model performance.
Phased training strategy
Adopts a phased training strategy from 512 to 8192, effectively supporting long-context learning.

Model Capabilities

Text embedding
Text re-ranking
Long-text processing
Masked language modeling

Use Cases

Information retrieval
Document retrieval
Used for semantic retrieval and ranking of long documents
Provides more accurate retrieval results in long-context scenarios
Natural language processing
Text representation learning
Generates high-quality text embedding representations
Can be used for feature extraction in downstream NLP tasks
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase