L

Llm2vec Meta Llama 3 8B Instruct Mntp

Developed by McGill-NLP
LLM2Vec is a simple solution for converting decoder-only large language models into text encoders, achieved by enabling bidirectional attention mechanisms, masked next-token prediction, and unsupervised contrastive learning.
Downloads 3,885
Release Time : 4/30/2024

Model Overview

This model transforms large language models into powerful text encoders through a three-step conversion process, supporting various tasks such as text embedding, information retrieval, and text classification.

Model Features

Bidirectional Attention Mechanism
By enabling bidirectional attention, decoder-only LLMs can better understand contextual information.
Masked Next-Token Prediction
Uses masked next-token prediction (MNTP) to enhance the model's text representation capabilities.
Unsupervised Contrastive Learning
Further improves text encoding quality through unsupervised contrastive learning.
Instruction-Aware Encoding
Supports text encoding with instruction prefixes, suitable for scenarios like retrieval-augmented generation.

Model Capabilities

Text embedding
Information retrieval
Text classification
Text clustering
Semantic similarity calculation
Feature extraction
Text reranking

Use Cases

Information Retrieval
QA System Retrieval
Uses instruction-encoded queries to retrieve relevant document passages.
Highly relevant document retrieval.
Text Analysis
Semantic Similarity Calculation
Calculates semantic similarity between different texts.
Accurate similarity scores.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase