L

Llama 3.2 400M Amharic

Developed by rasyosef
This is a streamlined version based on Meta's Llama-3.2-1B model, specifically pretrained for Amharic with 400 million parameters and a context length of 1024 tokens.
Downloads 310
Release Time : 11/26/2024

Model Overview

This model is a decoder transformer designed for Amharic text generation tasks without supervised fine-tuning.

Model Features

Amharic Optimization
Pretrained with 274 million Amharic text tokens, specifically optimized for Amharic text generation.
Streamlined Model
A compact version of the Llama-3.2-1B model with 400 million parameters, suitable for running on a single A100 40GB GPU.
Efficient Training
Completed pretraining in just 23 hours on a single A100 40GB GPU, achieving a validation perplexity of 41.3.

Model Capabilities

Amharic text generation
Long-text generation (1024 token context length)

Use Cases

Text generation
News Summary Generation
Generate news summaries based on Amharic news headlines
Produces coherent and contextually appropriate news content
Dialogue Systems
Used for reply generation in Amharic chatbots
Generates natural and fluent conversational responses
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase