N

Norbert2

Developed by ltg
NorBERT 3 is a series of Norwegian pre-trained language models, trained on a large-scale Norwegian corpus, supporting various natural language processing tasks.
Downloads 741
Release Time : 3/2/2022

Model Overview

NorBERT 3 is a Norwegian pre-trained language model based on the BERT architecture, specifically designed for Norwegian natural language processing tasks, suitable for scenarios such as text classification, named entity recognition, and question-answering systems.

Model Features

Large-scale corpus training
Trained on an ultra-large Norwegian corpus (C4 + NCC, approximately 15 billion tokens)
Whole word masking technique
Utilizes whole word masking to enhance the model's understanding of Norwegian
Multiple version options
Offers models with varying parameter sizes from ultra-lightweight to enhanced versions to meet different computational needs

Model Capabilities

Text understanding
Text generation
Fill-mask
Named entity recognition
Text classification

Use Cases

Text processing
Text completion
Automatically completes missing parts in Norwegian sentences
Example input: 'Nå ønsker de seg en [MASK] bolig.' Can predict suitable words like 'ny' (new)
Text classification
Classifies Norwegian text
Information extraction
Named entity recognition
Identifies entities such as person names and locations from Norwegian text
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase