D

Distilbert Dot Tas B B256 Msmarco

Developed by sebastian-hofstaetter
A dual-encoder dot-product scoring architecture based on DistilBert, trained on the MSMARCO-Passage dataset with balanced topic-aware sampling, suitable for dense retrieval and candidate set re-ranking
Downloads 3,188
Release Time : 3/2/2022

Model Overview

This model is an efficient dense passage retrieval system that excels in information retrieval tasks through knowledge distillation and topic-aware sampling training

Model Features

Balanced Topic-Aware Sampling
Utilizes the innovative TAS-B training method to optimize the sampling distribution of training data
Efficient Training
Training can be completed in just 48 hours on a single consumer-grade GPU
Dual Supervision Mechanism
Combines BERT_CAT pairwise scores and in-batch negative signals provided by the ColBERT model
Shared Encoding Architecture
Query and passage encoding share the same BERT layers, improving efficiency and reducing memory requirements

Model Capabilities

Dense Passage Retrieval
Candidate Set Re-ranking
Semantic Similarity Calculation

Use Cases

Information Retrieval
Search Engine Result Re-ranking
Semantic re-ranking of results returned by traditional retrieval systems
Achieves MRR@10 of 0.347 on MSMARCO-DEV
End-to-End Dense Retrieval
Directly used in vector index-based dense retrieval systems
Achieves Recall@1K of 0.843 on TREC-DL'19
Featured Recommended AI Models
ยฉ 2025AIbase