M

Meltemi 7B V1

Developed by ilsp
The first large-scale Greek foundational language model, based on the Mistral-7B architecture, enhanced with 40 billion tokens of Greek and English corpus to improve Greek language capabilities
Downloads 49
Release Time : 3/22/2024

Model Overview

The first Greek large language model trained by the Institute for Language and Speech Processing (ILSP) under the Athens Research and Innovation Center, supporting Greek and English with text generation capabilities

Model Features

Greek Language Optimization
Extended the Mistral-7B tokenizer to support Greek vocabulary, enhancing Greek language capabilities with 28.5 billion Greek tokens
Bilingual Capabilities
Maintained English proficiency while enhancing Greek capabilities, preventing catastrophic forgetting
Long Context Support
Supports context lengths of up to 8192 tokens
High-Quality Corpus
Trained on a rigorously cleaned and deduplicated 40 billion token corpus

Model Capabilities

Greek text generation
English text generation
Bilingual text processing

Use Cases

Education
Greek Language Learning Assistance
Helps learners understand and generate Greek content
Medical
Medical Q&A
Question-answering capabilities based on Greek medical exam datasets
Achieved 41.0% accuracy on the Greek Medical MCQA test set
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase