H

Herbert Base Cased

Developed by allegro
HerBERT is a Polish pre-trained language model based on the BERT architecture, trained using dynamic whole word masking and sentence structure objectives.
Downloads 84.18k
Release Time : 3/2/2022

Model Overview

HerBERT is an efficient Transformer model optimized for Polish, primarily used for natural language processing tasks such as text classification, question answering, and text generation.

Model Features

Polish optimization
Specially designed and trained for Polish, excelling in Polish language tasks.
Dynamic whole word masking
Uses dynamic whole word masking strategy for pre-training, improving the model's understanding of Polish.
Sentence structure objective
In addition to traditional MLM tasks, it is also trained using sentence structure objectives (SSO), enhancing the model's understanding of sentence structures.
Large-scale training data
Trained on over 8.5 billion Polish tokens, covering various text types.

Model Capabilities

Polish text understanding
Polish text generation
Polish text classification
Polish question answering systems

Use Cases

Natural language processing
Polish text classification
Can be used for tasks such as sentiment analysis and topic classification in Polish.
Polish question answering system
Build question answering systems for Polish content.
Polish text generation
Generate fluent Polish text.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase