D

Dpr Question Encoder Single Lfqa Wiki

Developed by vblagoje
A DPR-based question encoder model specifically designed for long-form QA (LFQA) tasks, optimized for retrieval performance through two-stage training
Downloads 588
Release Time : 3/2/2022

Model Overview

This model uses Transformer's pooled output as question representations, primarily for retrieving relevant answer passages from large-scale knowledge bases for long-form questions

Model Features

Two-stage Training Strategy
First stage fine-tunes with LFQA datasets, second stage introduces Wikipedia indexing to construct higher-quality training samples
Hard Negative Optimization
Enhances model discrimination through carefully designed negative sample selection strategy (cosine similarity range 0.55-0.65)
Cross-encoder Enhancement
Second stage uses SBert cross-encoder to score candidate answers and filter high-quality positive/negative samples

Model Capabilities

Question vector encoding
Semantic similarity calculation
Open-domain retrieval
Long-form QA support

Use Cases

Knowledge Retrieval Systems
Wikipedia QA System
Retrieves the most relevant answer passages from Wikipedia for complex questions
Can replace traditional keyword retrieval, providing semantically better-matched results
Educational Assistance
Learning Assistant
Helps students retrieve detailed explanations of complex concepts in long-form answers
Provides more comprehensive knowledge explanations than simple Q&A
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase