R

Roberta Base

Developed by FacebookAI
An English pre-trained model based on Transformer architecture, trained on massive text through masked language modeling objectives, supporting text feature extraction and downstream task fine-tuning
Downloads 9.3M
Release Time : 3/2/2022

Model Overview

A bidirectional Transformer model employing dynamic masking strategy, optimizing BERT's pre-training methods, suitable for NLP tasks like sequence classification and token classification

Model Features

Dynamic Masking Strategy
Adopts more efficient dynamic masking, learning more comprehensive contextual representations compared to static masking
Large-scale Training Data
Incorporates 5 datasets totaling 160GB of text, covering various genres including books, news, and encyclopedias
Optimized Training Configuration
Uses 8K batch size and 512 sequence length for 500K training steps, employing Adam optimizer with learning rate warm-up strategy

Model Capabilities

Text Feature Extraction
Masked Language Prediction
Sequence Classification
Token Classification
Question Answering

Use Cases

Text Understanding
Sentiment Analysis
Classify sentiment orientation of reviews/tweets
Achieves 94.8% accuracy on SST-2 dataset
Text Similarity Calculation
Measure semantic similarity between two texts
Scores 91.2 on STS-B dataset
Information Extraction
Named Entity Recognition
Identify entities like persons/locations/organizations from text
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase