T5 Small Ssm
Google's T5 model achieves closed-book QA through pre-training, capable of answering questions without relying on external knowledge sources
Downloads 88
Release Time : 3/2/2022
Model Overview
Based on the T5 architecture, this model is first pre-trained on the C4 dataset with denoising objectives, then additionally pre-trained on Wikipedia data using REALM's salient span masking objective, specifically designed for closed-book QA scenarios
Model Features
Closed-book QA capability
Can answer questions without external knowledge sources, with answers entirely derived from knowledge stored in the model's internal parameters
Dual pre-training strategy
First pre-trained on C4 dataset with standard denoising objectives, then knowledge-intensive pre-training on Wikipedia using REALM's salient span masking objectives
Scalability
Research shows model performance improves with scale, with larger versions performing comparably to open-domain QA systems
Model Capabilities
Closed-book QA
Knowledge retrieval
Text generation
Use Cases
Education
Knowledge QA system
Building automated QA systems that don't rely on external databases
Can achieve accuracy comparable to explicit retrieval systems
Research
Knowledge encapsulation research
Studying how much knowledge can be encapsulated in language model parameters
Validated the positive correlation between model scale and knowledge storage capacity
Featured Recommended AI Models
Š 2025AIbase