O

Openelm 3B Instruct

Developed by apple
OpenELM is a set of open-source and efficient language models. It adopts a hierarchical parameter allocation strategy to improve model accuracy and includes pre-trained and instruction-tuned versions with 270 million to 3 billion parameters.
Downloads 8,716
Release Time : 4/12/2024

Model Overview

OpenELM is a series of open-source and efficient language models launched by Apple. It optimizes the parameter distribution within Transformer layers through an innovative hierarchical parameter allocation strategy, significantly improving model efficiency. This series includes pre-trained models of various scales and instruction-fine-tuned versions, suitable for various natural language processing tasks.

Model Features

Hierarchical parameter allocation
Intelligently allocate parameters in each Transformer layer, significantly improving model efficiency compared to traditional uniform allocation
Complete open-source ecosystem
Provide a full set of toolchains from data preparation to training and evaluation, including the CoreNet training framework
Multiple scale options
Offer various model scales from 270 million to 3 billion parameters to meet different computing requirements
Instruction-tuned version
In addition to the basic pre-trained model, an instruction-fine-tuned version is also released to optimize dialogue interaction capabilities

Model Capabilities

Text generation
Language understanding
Question-answering system
Instruction following

Use Cases

Education and research
Academic Q&A
Based on excellent performance on scientific Q&A datasets such as SciQ, it is suitable for knowledge Q&A applications in educational scenarios
Achieved 92.7% accuracy on the SciQ dataset
Content generation
Creative writing
Utilize the model's text generation ability to assist in story creation
The example shows the generation of a story beginning with 'Once upon a time...'
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase