L

Long T5 Tglobal Large Pubmed 3k Booksum 16384 WIP15

Developed by pszemraj
A large-scale summarization model based on the Long-T5 architecture, specifically optimized for book and long document summarization tasks
Downloads 17
Release Time : 9/20/2022

Model Overview

This model is a variant based on the Long-T5 architecture, specifically designed for summarization tasks involving long documents and books. It is trained on the pubmed and booksum datasets and can handle input texts up to 16,384 tokens in length.

Model Features

Long Text Processing Capability
Supports processing input texts up to 16,384 tokens, making it suitable for book and long document summarization
Domain-Specific Optimization
Specifically trained on the pubmed medical literature and booksum book summarization datasets
Multi-Dataset Performance
Evaluated on multiple summarization datasets, including samsum, booksum, and gov_report

Model Capabilities

Long text summarization generation
Book content summarization
Technical document summarization
Government report summarization

Use Cases

Academic Research
Medical Literature Summarization
Generates concise summaries for medical literature such as pubmed
ROUGE-1 score 35.33 (booksum dataset)
Publishing Industry
Book Content Summarization
Generates chapter or full-book summaries for lengthy books
Average generation length 268 tokens (booksum dataset)
Government Document Processing
Government Report Summarization
Processes lengthy government reports and generates key-point summaries
ROUGE-1 score 37.04 (gov_report dataset)
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase