L

Legal Xlm Roberta Base

Developed by joelniklaus
A multilingual XLM-RoBERTa model pre-trained on legal data, supporting legal text processing in 24 European languages
Downloads 387
Release Time : 1/16/2023

Model Overview

This model is a further pre-trained version of XLM-RoBERTa base on multilingual legal corpora, specifically optimized for downstream tasks in the legal domain

Model Features

Legal domain optimization
Pre-trained specifically on 689GB of multilingual legal corpora, excelling in legal text processing
Multilingual support
Supports legal text processing in 24 European languages, including minority languages like Maltese and Irish
Long text processing capability
Optimized with window attention mechanism and 15% masking rate, suitable for processing lengthy legal texts

Model Capabilities

Legal text understanding
Multilingual text classification
Legal QA systems
Legal entity recognition

Use Cases

Legal text analysis
Legal document classification
Automatic classification of multilingual legal documents
Outperforms in LEXTREME benchmark tests
Legal QA system
Building cross-jurisdictional legal QA applications
Legal research assistance
Cross-jurisdictional legal provision comparison
Analyzing similarities and differences in legal provisions across countries
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase