O

Opensearch Neural Sparse Encoding Doc V3 Distill

Developed by opensearch-project
A document-level learned sparse retrieval model specifically designed for OpenSearch, optimized with distillation technology to support efficient document retrieval
Downloads 243
Release Time : 3/28/2025

Model Overview

This model encodes documents into 30522-dimensional sparse vectors, suitable for document retrieval tasks, with special optimizations for retrieval efficiency in OpenSearch

Model Features

Inference-free Retrieval
Document processing requires no inference computation, directly generating sparse vectors to significantly reduce computational costs
Efficient Retrieval
Average FLOPS only 1.8, significantly improving efficiency compared to previous generation models
Improved Relevance
Achieves average NDCG@10 of 0.517 on BEIR benchmark, outperforming previous document-specialized models
Large-scale Training
Trained on diverse QA and document datasets including MS MARCO, WikiAnswers, etc.

Model Capabilities

Document Retrieval
Sparse Vector Generation
Semantic Matching
Cross-domain Retrieval

Use Cases

Search Engine
OpenSearch Document Retrieval
Serves as OpenSearch's neural sparse retrieval component, providing efficient document search capabilities
Delivers better semantic matching effects compared to traditional retrieval methods
QA Systems
QA Pair Retrieval
Retrieves the most relevant QA pairs from knowledge base for user questions
Performs well on QA datasets like NQ
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase