Norbert2
NorBERT 3 is a series of Norwegian pre-trained language models, trained on a large-scale Norwegian corpus, supporting various natural language processing tasks.
Downloads 741
Release Time : 3/2/2022
Model Overview
NorBERT 3 is a Norwegian pre-trained language model based on the BERT architecture, specifically designed for Norwegian natural language processing tasks, suitable for scenarios such as text classification, named entity recognition, and question-answering systems.
Model Features
Large-scale corpus training
Trained on an ultra-large Norwegian corpus (C4 + NCC, approximately 15 billion tokens)
Whole word masking technique
Utilizes whole word masking to enhance the model's understanding of Norwegian
Multiple version options
Offers models with varying parameter sizes from ultra-lightweight to enhanced versions to meet different computational needs
Model Capabilities
Text understanding
Text generation
Fill-mask
Named entity recognition
Text classification
Use Cases
Text processing
Text completion
Automatically completes missing parts in Norwegian sentences
Example input: 'Nå ønsker de seg en [MASK] bolig.' Can predict suitable words like 'ny' (new)
Text classification
Classifies Norwegian text
Information extraction
Named entity recognition
Identifies entities such as person names and locations from Norwegian text
Featured Recommended AI Models