D

Dense Encoder Msmarco Distilbert Word2vec256k MLM 785k Emb Updated

Developed by vocab-transformers
A DistilBERT model with word2vec-initialized vocabulary, optimized for sentence similarity tasks and trained on the MS MARCO dataset
Downloads 33
Release Time : 3/2/2022

Model Overview

This model uses a 256k vocabulary initialized by word2vec, pre-trained for 785k steps with MLM, and fine-tuned on the MS MARCO dataset using MarginMSELoss, suitable for sentence embedding and semantic search tasks

Model Features

Large Vocabulary
256k vocabulary initialized by word2vec, covering a broader semantic range
Efficient Training
Based on DistilBERT architecture, reducing computational resource requirements while maintaining performance
Optimized Loss Function
Trained with MarginMSELoss to optimize performance on sentence similarity tasks

Model Capabilities

Sentence Embedding Generation
Semantic Similarity Calculation
Text Clustering
Information Retrieval

Use Cases

Information Retrieval
Document Search
Convert queries and documents into vectors for similarity matching
Achieved MRR@10 of 35.20 on the MS MARCO development set
Question Answering Systems
Question Matching
Identify semantically similar question pairs
Achieved nDCG@10 of 67.61 and 69.62 on TREC-DL 2019/2020 respectively
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase