🚀 LAION LeoLM: Linguistically Enhanced Open Language Model
Meet LeoLM, the first open and commercially available German Foundation Language Model built on Llama-2. It extends Llama-2's capabilities into German through continued pretraining on a large corpus of German - language and mostly locality - specific text.
🚀 Quick Start
LeoLM is the first open and commercially available German Foundation Language Model based on Llama - 2. By continuing pretraining on a large German - language corpus, it extends Llama - 2's capabilities to German. Thanks to a compute grant at HessianAI's supercomputer 42, we release two 8k context - length foundation models, [LeoLM/leo - hessianai - 7b
](https://huggingface.co/LeoLM/leo - hessianai - 7b) and [LeoLM/leo - hessianai - 13b
](https://huggingface.co/LeoLM/leo - hessianai - 13b), under the [Llama - 2 community license](https://huggingface.co/meta - llama/Llama - 2 - 70b/raw/main/LICENSE.txt) (70b is also coming soon!). We hope this release will bring new opportunities to German open - source and commercial LLM research and accelerate adoption. Read our blog post or paper (preprint coming soon) for more details.
A project by Björn Plüster and Christoph Schuhmann in collaboration with LAION and HessianAI.
✨ Features
- Extends Llama - 2's capabilities to the German language through continued pretraining on a large German - language corpus.
- Released two 8k context - length foundation models, facilitating German open - source and commercial LLM research.
📦 Installation
Install Direct Dependencies
pip install transformers torch sentencepiece
Install Dependencies for Faster Inference with Flash - Attention2
pip install packaging ninja
pip install flash-attn==v2.1.1 --no-build-isolation
pip install git+https://github.com/HazyResearch/flash-attention.git@v2.1.1#subdirectory=csrc/rotary
💻 Usage Examples
Basic Usage
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch
model = AutoModelForCausalLM.from_pretrained(
model="LeoLM/leo-hessianai-13b",
device_map="auto",
torch_dtype=torch.float16,
trust_remote_code=True
)
📚 Documentation
Model Details
Property |
Details |
Finetuned from |
[meta - llama/Llama - 2 - 13b - hf](https://huggingface.co/meta - llama/Llama - 2 - 13b - hf) |
Model Type |
Causal decoder - only transformer language model |
Language |
English and German |
License |
[LLAMA 2 COMMUNITY LICENSE AGREEMENT](https://huggingface.co/meta - llama/Llama - 2 - 70b/raw/main/LICENSE.txt) |
Contact |
LAION Discord or Björn Plüster |
Training parameters

Benchmarks

📄 License
This model is released under the [LLAMA 2 COMMUNITY LICENSE AGREEMENT](https://huggingface.co/meta - llama/Llama - 2 - 70b/raw/main/LICENSE.txt).