Gte En Mlm Large
Large English text encoder in the GTE-v1.5 series, supporting context lengths up to 8192, built on an improved BERT architecture.
Downloads 171
Release Time : 8/6/2024
Model Overview
This model is a general-purpose text encoder developed by Alibaba Group's Institute of Intelligent Computing, primarily used for English text embedding representation and re-ranking tasks, supporting long-context processing.
Model Features
Long-context support
Supports context lengths up to 8192, suitable for processing long documents and complex texts.
Improved BERT architecture
Enhanced architecture combining RoPE and GLU, improving model performance.
Phased training strategy
Adopts a phased training strategy from 512 to 8192, effectively supporting long-context learning.
Model Capabilities
Text embedding
Text re-ranking
Long-text processing
Masked language modeling
Use Cases
Information retrieval
Document retrieval
Used for semantic retrieval and ranking of long documents
Provides more accurate retrieval results in long-context scenarios
Natural language processing
Text representation learning
Generates high-quality text embedding representations
Can be used for feature extraction in downstream NLP tasks
Featured Recommended AI Models
Š 2025AIbase