L

Lsg Bart Base 4096 Wcep

Developed by ccdv
Long text summarization model based on LSG-BART architecture, fine-tuned on the WCEP-10 dataset, supports processing long sequences of up to 4096 tokens
Downloads 27
Release Time : 5/25/2022

Model Overview

This model employs a local-sparse-global attention mechanism to handle long sequences, making it suitable for long text summarization tasks. It is based on a modified BART-base architecture and supports efficient processing of long documents.

Model Features

Long Sequence Processing Capability
Supports processing long texts up to 4096 tokens through the local-sparse-global attention mechanism for efficient computation
Multiple Sparse Attention Modes
Provides various sparse attention configuration options including local, pooling, strided, block-strided, normalized, and LSH
Resource Efficiency Optimization
Allows balancing performance and resource consumption by adjusting block size and sparsity

Model Capabilities

Long Text Summarization
Multi-Document Summarization
English Text Processing

Use Cases

Text Summarization
News Article Summarization
Generates concise summaries for long news articles
Achieves a ROUGE-1 score of 46.02 on the WCEP-10 test set
Research Report Condensation
Extracts key information from lengthy research reports
Supports processing long documents up to 4096 tokens
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase