W

Wangchanberta Base Att Spm Uncased

Developed by airesearch
A RoBERTa BASE model pre-trained on 78.5GB of Thai text, suitable for masked language modeling and text classification tasks
Downloads 34.38k
Release Time : 3/2/2022

Model Overview

This model is a Thai pre-trained model based on the RoBERTa architecture, primarily used for masked language modeling prediction and text classification tasks. Supports various downstream NLP applications.

Model Features

Large-scale Thai pre-training
Trained on 78.5GB of Thai text data with excellent Thai language understanding capabilities
Multi-task support
Supports various NLP tasks including masked language modeling, text classification, and sequence labeling
Optimized RoBERTa architecture
Utilizes an improved RoBERTa architecture with enhanced contextual understanding capabilities

Model Capabilities

Thai text understanding
Masked language prediction
Sentiment analysis
Text classification
Sequence labeling

Use Cases

Social media analysis
Sentiment analysis
Analyze sentiment tendencies in Thai social media posts
Can classify into four sentiment categories: positive/neutral/negative/question
Review analysis
Review rating prediction
Predict user review ratings (1-5 stars)
Performs well on the Wongnai review dataset
News classification
Multi-label topic classification
Classify Thai news into 12 topic categories
Achieves multi-label classification on the Prachathai67k dataset
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase