T

Twhin Bert Base

Developed by Twitter
TwHIN-BERT is a multilingual tweet language model trained on 7 billion tweets, optimized with text self-supervised learning and social interaction data.
Downloads 3,789
Release Time : 10/18/2022

Model Overview

TwHIN-BERT is a novel multilingual tweet language model trained using rich social interaction data from Twitter's Heterogeneous Information Network (TwHIN) combined with text self-supervised learning, suitable for various NLP and recommendation tasks.

Model Features

Multilingual Support
Supports over 100 languages, particularly suitable for processing multilingual tweet content.
Social-Enhanced Training
Optimized with social interaction data from Twitter's Heterogeneous Information Network (TwHIN), enhancing performance in social recommendation tasks.
Dual Training Objectives
Trained using both text self-supervised learning (e.g., masked language modeling) and social objective functions.

Model Capabilities

Multilingual Text Understanding
Social Recommendation
Text Classification
User Behavior Prediction

Use Cases

Natural Language Processing
Multilingual Text Classification
Classify tweet content with support for multiple languages.
Outperforms peer models in semantic understanding tasks
Social Recommendation
User-Tweet Interaction Prediction
Predict user interactions with tweets.
Excels in social recommendation tasks
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase