N

Nystromformer 512

Developed by uw-madison
An efficient Transformer model optimized with the Nyström method for handling long-sequence tasks
Downloads 1,570
Release Time : 3/2/2022

Model Overview

Nyströmformer is an improved Transformer model with a self-attention mechanism that approximates standard self-attention using the Nyström method, significantly reducing computational complexity and enabling efficient processing of long-sequence tasks.

Model Features

Efficient Self-Attention Mechanism
Uses the Nyström method to approximate standard self-attention, reducing complexity from O(n²) to O(n)
Long Sequence Processing Capability
Particularly suitable for handling long-sequence tasks involving thousands of tokens
Superior Performance
Demonstrates excellent performance in the GLUE benchmark and Long-Range Arena (LRA) benchmark

Model Capabilities

Text Infilling
Language Modeling
Long Text Processing

Use Cases

Natural Language Processing
Text Completion
Predicts and fills in missing parts of text
As shown in the example, it can accurately predict 'Paris is the capital of France'
Long Document Analysis
Processes and analyzes lengthy document content
Due to the optimized attention mechanism, it can effectively handle long-sequence inputs
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase