O

Openmoe Base

Developed by OrionZheng
OpenMoE-Base is a Mixture of Experts (MoE) base model for debugging purposes, trained on only 128 billion tokens. As part of the OpenMoE project, it aims to advance the open-source MoE community.
Downloads 73
Release Time : 12/8/2023

Model Overview

OpenMoE-Base is a large language model based on the Mixture of Experts (MoE) architecture, primarily used to validate the effectiveness of the model architecture. The project is committed to fully sharing resources such as training data, strategies, architecture, and weights.

Model Features

Open-Source MoE Architecture
Fully open-source implementation of the Mixture of Experts model, promoting research in the MoE field
Complete Training Resources
Publicly available training data, strategies, architecture, and all other resources
Debugging and Validation Purpose
Primarily used to validate the effectiveness of the model architecture, performance may be suboptimal

Model Capabilities

Text Generation
Language Understanding

Use Cases

Research and Development
MoE Architecture Research
Used to study the training dynamics and architectural characteristics of Mixture of Experts models
Validated the effectiveness of the MoE architecture
Debugging Benchmark
Serves as a debugging benchmark for the development of larger-scale MoE models
Laid the foundation for the development of 8B and 34B versions
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase