L

Lola V1

Developed by dice-research
LOLA is an ultra-large-scale multilingual large model based on the sparse Mixture-of-Experts (MoE) Transformer architecture, supporting over 160 languages, with competitive advantages in natural language generation and understanding tasks.
Downloads 867
Release Time : 4/2/2024

Model Overview

LOLA is an open-source multilingual large model that adopts a GPT2-style decoder-only architecture combined with sparse Mixture-of-Experts technology, supporting text generation tasks in over 160 languages.

Model Features

Multilingual support
Supports over 160 languages, excelling in multilingual natural language processing tasks
Mixture-of-Experts architecture
Employs a sparse Mixture-of-Experts (MoE) architecture with 16 experts, enhancing model performance while maintaining efficiency
Open-source and reproducible
Fully open-source, promoting research reproducibility and laying the foundation for future studies
Computationally efficient
Optimizes computational resource usage through expert routing mechanisms, activating only a subset of parameters per token

Model Capabilities

Multilingual text generation
Causal language modeling
Natural language understanding

Use Cases

Text generation
Multilingual text completion
Generates coherent subsequent content based on given text fragments
Example: Input 'The quick brown fox' outputs 'The quick brown fox jumps over the lazy dog.'
Language research
Cross-linguistic pattern analysis
Studies implicit linguistic genealogical patterns across different languages
The model reveals how expert routing mechanisms leverage implicit linguistic genealogical patterns
Featured Recommended AI Models
ยฉ 2025AIbase