Mallam 1.1B 4096
A 1.1B parameter pre-trained model based on Malay text, using Mistral architecture, supporting 4096 context length
Downloads 201
Release Time : 11/27/2023
Model Overview
This is a 1.1B parameter large language model specifically optimized for Malay, pre-trained from scratch on 90 billion Malay text tokens, suitable for Malay text generation and comprehension tasks
Model Features
Malay language optimization
Specifically trained and optimized for Malay text
Long context support
Supports long context processing capability of 4096 tokens
Efficient training
Efficiently completed training using a Ray cluster on 5 nodes (each with 4ÃA100 80GB)
Model Capabilities
Malay text generation
Long text comprehension
Language model reasoning
Use Cases
Text generation
Malay content creation
Generate Malay articles, stories, or other creative content
Dialogue systems
Build Malay chatbots or virtual assistants
Education
Language learning assistance
Help users learning Malay to practice and understand the language
Featured Recommended AI Models
Š 2025AIbase