O

Openelm 270M

Developed by apple
OpenELM is a set of open-source efficient language models that adopt a hierarchical scaling strategy to efficiently allocate parameters in each layer of the Transformer model, improving accuracy.
Downloads 4,719
Release Time : 4/12/2024

Model Overview

The OpenELM series includes pre-trained and instruction-tuned models with 270M, 450M, 1.1B, and 3B parameters, providing a complete training framework and evaluation process.

Model Features

Hierarchical Scaling Strategy
Efficiently allocates parameters in each layer of the Transformer model to enhance accuracy.
Complete Open-source Framework
Provides a full pipeline from data preparation, training, fine-tuning to evaluation, promoting open research.
Multiple Size Options
Offers model choices with 270M, 450M, 1.1B, and 3B parameter sizes.

Model Capabilities

Text Generation
Instruction Following
Zero-shot Learning

Use Cases

Natural Language Processing
Question Answering Systems
Can be used to build knowledge-based question answering systems.
Achieved 84.70% accuracy on the SciQ dataset
Text Completion
Generates coherent text content based on prompts.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase