Bertweet Large
BERTweet is the first large-scale pretrained language model specifically designed for English tweets, trained on the RoBERTa architecture, suitable for social media text analysis.
Downloads 2,853
Release Time : 3/2/2022
Model Overview
BERTweet is a pretrained language model optimized for English tweets, capable of handling language features and informal expressions unique to social media.
Model Features
Social Media Optimization
Specifically trained for tweet characteristics, effectively handling informal language, abbreviations, and emojis in social media.
Large-scale Training Data
Trained on 850 million English tweets (16 billion tokens), including COVID-19 related corpus.
Multi-task Support
Excellent performance on multiple NLP tasks such as POS tagging, Named Entity Recognition, Sentiment Analysis, and Sarcasm Detection.
Model Capabilities
Text Understanding
Sentiment Analysis
Named Entity Recognition
POS Tagging
Sarcasm Detection
Use Cases
Social Media Analysis
Public Opinion Monitoring
Analyzing public sentiment and opinion trends in tweets.
Outperforms general-purpose language models in sentiment analysis tasks.
Event Detection
Tracking trending events and personalities through Named Entity Recognition.
Accurately identifies informal naming expressions unique to social media.
Content Moderation
Sarcasm Content Identification
Detecting sarcastic and ironic content in tweets.
Achieves significant results in sarcasm detection tasks.
Featured Recommended AI Models