X

Xlm Roberta Longformer Base 4096

Developed by Peltarion
Extended XLM-RoBERTa model supporting sequences up to 4096 tokens, suitable for multilingual tasks
Downloads 64
Release Time : 3/2/2022

Model Overview

The XLM-R Long Sequence Model is an extended version of XLM-RoBERTa that supports longer sequence processing (original version only supports 512 tokens) through special pre-training. It performs exceptionally well on multilingual QA tasks, especially for low-resource language scenarios.

Model Features

Extended Context Support
Supports processing sequences up to 4096 tokens (original XLM-R only supports 512 tokens)
Low-resource Language Optimization
No need for separate pre-training per language, especially suitable for low-resource languages like Swedish
Efficient Training Scheme
Uses gradient accumulation (64 steps) and 16-bit precision training to reduce GPU memory requirements

Model Capabilities

Long-text understanding
Multilingual QA
Cross-lingual transfer learning

Use Cases

QA Systems
Multilingual Long-document QA
Handles cross-lingual QA tasks involving long contexts
Text Understanding
Low-resource Language Document Analysis
Semantic analysis of long documents in low-resource languages like Swedish
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase