Long T5 Tglobal Base 16384 Booksum V11 Big Patent V2
A long-text summarization model based on the T5 architecture, capable of processing inputs up to 16,384 tokens, suitable for book and technical document summarization tasks.
Downloads 21
Release Time : 7/31/2022
Model Overview
This model is an optimized long-text summarization model based on the T5 architecture, specifically trained for book and technical document summarization tasks. It can handle input sequences up to 16,384 tokens, making it suitable for generating concise summaries of book chapters, technical patents, and other lengthy documents.
Model Features
Ultra-Long Context Handling
Supports processing input sequences up to 16,384 tokens, ideal for summarizing lengthy documents like books.
Multi-Domain Adaptation
Trained on both book summarization (kmfoda/booksum) and technical patent (big_patent) datasets.
Efficient Attention Mechanism
Utilizes the TGlobal attention variant for optimized long-sequence processing efficiency.
Model Capabilities
Long-text summarization generation
Book chapter summarization
Technical document summarization
Content condensation
Use Cases
Publishing & Education
Book Chapter Summarization
Generates concise summaries for book chapters.
Achieved ROUGE-1 score of 23.14 on the booksum dataset.
Technical Document Processing
Patent Document Summarization
Generates key content summaries for technical patent documents.
Optimized through training on the big_patent dataset.
Featured Recommended AI Models
Š 2025AIbase