M

M2 BERT 2k Retrieval Encoder V1

Developed by hazyresearch
80M-parameter M2-BERT-2k model checkpoint, specifically designed for long-context retrieval tasks, supporting a context length of 2048 tokens.
Downloads 80
Release Time : 5/22/2024

Model Overview

M2-BERT is an improved model based on the BERT architecture, specifically optimized for long-context retrieval tasks. It can generate 768-dimensional embedding vectors, suitable for scenarios such as information retrieval.

Model Features

Long Context Support
Supports processing of long contexts up to 2048 tokens, ideal for long-document retrieval tasks
Efficient Retrieval Embeddings
Generates high-quality 768-dimensional embedding vectors optimized for retrieval performance
Lightweight Architecture
Lightweight design with only 80M parameters, reducing computational resource requirements while maintaining performance

Model Capabilities

Text Embedding Generation
Long Text Processing
Information Retrieval

Use Cases

Information Retrieval
Document Retrieval
Using model-generated embeddings for similar document retrieval
Effectively handles documents up to 2048 tokens in length
Semantic Search
Content search system based on semantic similarity
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase