M

M2 Bert 80M 32k Retrieval

Developed by togethercomputer
This is an 80M parameter M2-BERT pre-trained model, supporting sequences up to 32,768 tokens in length, specifically optimized for long-context retrieval tasks.
Downloads 1,274
Release Time : 11/4/2023

Model Overview

A BERT variant model based on the Monarch Mixer architecture, fine-tuned for long-text retrieval tasks, capable of generating high-quality text embeddings.

Model Features

Ultra-Long Context Processing
Supports sequences up to 32,768 tokens, making it suitable for long-document retrieval tasks.
Efficient Architecture
Utilizes the Monarch Mixer sub-quadratic architecture to maintain performance while improving computational efficiency.
Retrieval Optimization
Fine-tuned specifically for retrieval tasks, generating high-quality 768-dimensional text embeddings.

Model Capabilities

Long Text Similarity Calculation
Semantic Retrieval
Text Embedding Generation

Use Cases

Information Retrieval
Long Document Retrieval
Quickly find relevant content from a large collection of long documents.
Capable of effectively processing documents up to 32k tokens in length.
Semantic Search
Document search based on semantics rather than keywords.
Generates high-quality semantic embedding vectors.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase