H

Herbert Large Cased

Developed by allegro
HerBERT is a Polish pre-trained language model based on the BERT architecture, trained using dynamic whole word masking and sentence structure objectives.
Downloads 1,272
Release Time : 3/2/2022

Model Overview

HerBERT is an efficient Polish pre-trained language model based on the BERT architecture, suitable for various natural language processing tasks.

Model Features

Dynamic whole word masking
Trained using masked language modeling with dynamic whole word masking, enhancing the model's language understanding capabilities.
Sentence structure objective
Incorporates sentence structure objectives (SSO) during training to improve the model's understanding of sentence structures.
Large-scale training corpus
Trained on six Polish corpora, covering a wide range of text types and domains.
Efficient tokenizer
Uses character-level byte pair encoding (CharBPETokenizer) to convert text into 50K subword units, improving processing efficiency.

Model Capabilities

Polish text understanding
Polish text generation
Masked language modeling

Use Cases

Natural language processing
Text classification
Used for Polish text classification tasks such as sentiment analysis and topic classification.
Named entity recognition
Identifies named entities in Polish text, such as person names, locations, and organization names.
Machine translation
Serves as a component in Polish machine translation systems to improve translation quality.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase