R

Roberta Large

Developed by FacebookAI
A large English language model pre-trained with masked language modeling objectives, using improved BERT training methods
Downloads 19.4M
Release Time : 3/2/2022

Model Overview

RoBERTa is a transformer model pre-trained through self-supervised learning on massive English corpora, primarily used for text feature extraction and downstream task fine-tuning

Model Features

Dynamic Masking Strategy
Unlike BERT, it employs a dynamic masking mechanism that generates different masking patterns in each iteration
Large-scale Training Data
Trained on 160GB of text data including books, Wikipedia, news, and various other sources
Optimized Training Process
Uses larger batches (8K) and longer sequences (512), trained for 500K steps

Model Capabilities

Text Feature Extraction
Masked Word Prediction
Downstream Task Fine-tuning

Use Cases

Natural Language Processing
Sequence Classification
Used for sentiment analysis, text classification tasks
Achieved 96.4 on SST-2 sentiment analysis task
Question Answering System
Serves as the base model for question answering systems
Achieved 94.7 on QNLI task
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase