L

Lsg Bart Base 16384 Pubmed

Developed by ccdv
A long-sequence text summarization model based on the BART architecture, specifically fine-tuned for the PubMed scientific paper dataset, capable of processing input sequences up to 16,384 tokens in length
Downloads 22
Release Time : 5/9/2022

Model Overview

This model employs a local-sparse-global attention mechanism for long-sequence text summarization tasks, particularly suitable for generating summaries of lengthy texts such as scientific papers

Model Features

Long Sequence Processing Capability
Capable of processing input sequences up to 16,384 tokens, making it particularly suitable for long document summarization
Efficient Attention Mechanism
Utilizes a local-sparse-global attention mechanism to enhance long-sequence processing efficiency while maintaining performance
Scientific Paper Optimization
Specifically fine-tuned for the PubMed scientific paper dataset, ideal for academic text summarization

Model Capabilities

Long Text Summarization Generation
Scientific Paper Content Extraction
English Text Processing

Use Cases

Academic Research
Automatic Summarization of Scientific Papers
Generates concise and accurate summaries for lengthy research papers
Achieves a ROUGE-1 score of 48.32 on the PubMed dataset
Literature Review Assistance
Helps researchers quickly grasp the core content of multiple papers
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase