W

Wangchanberta Base Wiki Newmm

Developed by airesearch
A RoBERTa BASE model pretrained on Thai Wikipedia, suitable for Thai text processing tasks
Downloads 115
Release Time : 3/2/2022

Model Overview

This model is a RoBERTa BASE architecture model pretrained on the Thai Wikipedia corpus, primarily used for masked language modeling tasks in Thai, and also applicable for text classification and token classification tasks.

Model Features

Thai Language Optimization
Specifically pretrained and optimized for Thai text
Multi-task Support
Supports various downstream tasks including text classification and named entity recognition
Large-scale Pretraining
Pretrained on a large-scale Thai Wikipedia corpus

Model Capabilities

Masked Language Modeling
Text Classification
Named Entity Recognition
Part-of-Speech Tagging

Use Cases

Sentiment Analysis
Social Media Sentiment Analysis
Analyze sentiment tendencies in social media posts and tweets
Supports 4 sentiment categories (Positive, Neutral, Negative, Question)
Review Analysis
User Review Rating Prediction
Predict star ratings (1-5 stars) for user reviews
News Classification
News Topic Classification
Multi-label topic classification for news articles
Supports 12 topic labels
Information Extraction
Named Entity Recognition
Identify named entities from text
Supports 13 named entity types
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase