Bert Small2bert Small Finetuned Cnn Daily Mail Summarization
This is an encoder-decoder model based on the BERT-small architecture, specifically fine-tuned for summarization tasks on the CNN/Dailymail dataset.
Downloads 92
Release Time : 3/2/2022
Model Overview
The model employs a warm-start BERT2BERT small architecture, fine-tuned on the CNN/Dailymail dataset to generate high-quality text summaries.
Model Features
Warm-start training
The model is warm-started based on the pre-trained BERT-small architecture, improving training efficiency and performance.
Efficient summarization
Optimized for news text, capable of generating concise and accurate summaries.
Lightweight architecture
Uses a small BERT architecture, reducing computational resource requirements while maintaining performance.
Model Capabilities
Text summarization generation
News content condensation
Long text compression
Use Cases
News processing
Automatic news summarization
Generates brief summaries for long news articles
ROUGE-2 score of 17.37 on CNN/Dailymail test set
Content analysis
Key information extraction from documents
Extracts core information from long documents
Featured Recommended AI Models