O

Openelm 3B

Developed by apple
OpenELM is a set of open-source efficient language models that employ a hierarchical scaling strategy to optimize parameter allocation and improve model accuracy. It includes four parameter scales: 270M, 450M, 1.1B, and 3B, offering both pre-trained and instruction-tuned versions.
Downloads 1,436
Release Time : 4/12/2024

Model Overview

An open-source series of efficient language models that achieve superior performance through hierarchical parameter allocation strategies, suitable for various natural language processing tasks.

Model Features

Hierarchical Parameter Allocation
Intelligently allocates parameters across Transformer layers, significantly improving model efficiency compared to traditional uniform allocation.
Full-process Open Source
Completely discloses the entire framework from data preparation to model training, fine-tuning, and evaluation.
Multiple Size Options
Offers models ranging from 270 million to 3 billion parameters to accommodate diverse application scenarios.

Model Capabilities

Text Generation
Zero-shot Learning
Instruction Following
Language Understanding

Use Cases

Education
Scientific Q&A
High-accuracy scientific question answering based on datasets like SciQ
The 3B model achieved 92.7% accuracy on the SciQ test set.
Content Creation
Story Generation
Generates coherent narrative texts based on prompts
Supports Chinese prompts such as 'Once upon a time' for story beginnings.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase