L

Lsg Bart Base 4096 Pubmed

Developed by ccdv
A long-sequence processing model based on LSG attention mechanism, fine-tuned specifically for scientific paper summarization tasks
Downloads 21
Release Time : 5/9/2022

Model Overview

This model is an improved version of BART-base, employing local-sparse-global attention mechanism to handle long sequence inputs. Fine-tuned on the PubMed scientific paper dataset, it is suitable for long-text summarization tasks.

Model Features

Long Sequence Processing Capability
Supports input sequences up to 4096 tokens, efficiently processing long texts using local-sparse-global attention mechanism
Multiple Attention Modes
Offers various sparse attention modes including local, pooled, strided, block-strided, normalized, and LSH
Scientific Paper Optimization
Specially fine-tuned on the PubMed scientific paper dataset, ideal for academic text summarization

Model Capabilities

Long Text Processing
Scientific Paper Summarization
Sequence-to-Sequence Transformation

Use Cases

Academic Research
Automatic Scientific Paper Summarization
Generates concise and accurate summaries for lengthy research papers
Achieves ROUGE-1 score of 47.37 on PubMed test set
Literature Processing
Medical Literature Summarization
Processes lengthy medical research literature to extract key information
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase