L

Lsg Bart Large 4096

Developed by ccdv
The LSG model is an improved long-sequence processing model based on BART-large, utilizing local + sparse + global attention mechanisms for efficient handling of long-text tasks
Downloads 15
Release Time : 3/2/2022

Model Overview

This model is optimized for encoder-decoder tasks, capable of efficiently processing long-sequence inputs with faster speed and higher efficiency compared to traditional long-sequence models

Model Features

Efficient Long-Sequence Processing
Utilizes local + sparse + global attention mechanisms (LSG), significantly improving long-text processing efficiency
Adaptive Sequence Length
Supports automatic padding of sequence length to integer multiples of chunk size, ensuring processing stability
Multi-Mode Sparse Selection
Provides 6 sparse selection modes (e.g., BOS pooling, LSH clustering, etc.) to adapt to different task requirements
Compatible with Original Architecture
Maintains the same parameter scale and layer count as BART-large, sharing the same tokenizer

Model Capabilities

Long Text Summarization
Sequence-to-Sequence Conversion
Efficient Processing of 4096-Length Inputs
Text Classification

Use Cases

Text Summarization
Automatic Summarization of Long Documents
Generates precise summaries for research papers, long articles, and other lengthy texts
Significantly improved processing speed compared to traditional models
Text Processing
Long Text Classification
Performs classification tasks on lengthy documents
Maintains high accuracy while reducing memory consumption
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase