L

Long T5 Tglobal Large Pubmed 3k Booksum 16384 WIP

Developed by pszemraj
A large-scale summarization model based on the Long-T5 architecture, specifically optimized for long-document summarization tasks, supporting a context length of 16,384 tokens.
Downloads 65
Release Time : 6/25/2022

Model Overview

This model is an improved long-text summarization generation model based on the T5 architecture, specifically optimized for book summarization and long-document summarization tasks. It was trained on the PubMed and BookSum datasets and can process input texts up to 16,384 tokens in length.

Model Features

Long Text Processing Capability
Supports ultra-long context processing of 16,384 tokens, suitable for book and long-document summarization
Multi-domain Adaptability
Trained on medical literature (PubMed) and book summarization (BookSum) datasets, applicable to various professional fields
High-performance Summarization
Achieves a ROUGE-1 score of 35.9969 on the BookSum test set, demonstrating excellent performance

Model Capabilities

Long text summarization generation
Book content summarization
Professional literature summarization
Multi-paragraph text condensation

Use Cases

Academic Research
Medical Literature Summarization
Automatically generates concise summaries of PubMed medical research papers
Maintains high information retention in professional fields
Publishing Industry
Book Content Summarization
Generates content summaries and chapter outlines for long books
Performs excellently on the BookSum dataset
Business Analysis
Long Report Condensation
Condenses business reports, market analyses, and other long documents into executive summaries
Retains key information and data points
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase