L

Longformer Zh

Developed by ValkyriaLenneth
A PyTorch-based Chinese Longformer model capable of processing document sequences up to 4096 characters with linear complexity, suitable for Chinese long-text tasks.
Downloads 418
Release Time : 3/2/2022

Model Overview

This model integrates local window attention with task-oriented global attention, perfectly replacing standard self-attention modules, especially suitable for Chinese long-text tasks.

Model Features

Linear Complexity for Long Text Processing
Compared to Transformer's O(n^2) complexity, it can process document sequences up to 4096 characters with linear complexity.
Hybrid Attention Mechanism
Integrates local window attention with task-oriented global attention, perfectly replacing standard self-attention modules.
Whole Word Masking Mechanism
Introduces Whole Word Masking (WWM) mechanism adapted to Chinese characteristics, reportedly the first open-source PyTorch implementation of Chinese WWM.

Model Capabilities

Long Text Processing
Text Classification
Reading Comprehension
Coreference Resolution
Sentiment Analysis

Use Cases

Sentiment Analysis
CCF Sentiment Analysis
Used for Chinese text sentiment classification tasks
Development set F1 reached 80.51, comparable to Roberta-mid
Reading Comprehension
Chinese Reading Comprehension (CMRC)
Used for Chinese reading comprehension tasks
F1:86.15, EM:66.84, outperforming Bert baseline
Coreference Resolution
Coreference Resolution Task
Used for Chinese coreference resolution tasks
Conll-F1:67.81, outperforming Bert and Roberta
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase