M

Minerva 1B Base V1.0

Developed by sapienzanlp
Minerva-1B-base-v1.0 is a 1-billion-parameter Italian-English bilingual large language model jointly developed by Sapienza NLP with FAIR and CINECA, trained on 200 billion tokens (100 billion each for Italian/English)
Downloads 551
Release Time : 4/19/2024

Model Overview

The first member of a large language model family fully pre-trained on Italian, a truly open Italian-English bilingual foundational model without alignment processing

Model Features

Balanced bilingual training
Uses balanced training data of 100 billion Italian tokens and 100 billion English tokens
Efficient tokenizer
Italian tokenization efficiency of 1.39 (CulturaX test), outperforming similar models
Long context support
Supports context lengths up to 16,384 tokens

Model Capabilities

Italian text generation
English text generation
Multitasking

Use Cases

Education
Italian learning assistance
Generates Italian learning materials and exercises
Content creation
Bilingual content generation
Automatically generates Italian-English bilingual articles or marketing content
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase