M

Mpt 30b

Developed by mosaicml
MPT-30B is an open-source large language model trained by MosaicML, based on a decoder-only Transformer architecture, pre-trained on 1 trillion English text and code tokens, supporting an 8k context window and efficient inference.
Downloads 2,021
Release Time : 6/20/2023

Model Overview

MPT-30B is part of Mosaic's Pretrained Transformer series, featuring an optimized Transformer architecture for efficient training and inference, with strong text generation and coding capabilities, suitable for various natural language processing tasks.

Model Features

Extended context support
Supports an 8k token context window, extendable to longer sequences via ALiBi technology
Efficient inference
Fast inference enabled by FlashAttention and FasterTransformer, deployable on a single GPU
Business-friendly license
Uses Apache-2.0 license, permitting commercial use (unlike restrictive licenses like LLaMA)
Massive training data
Trained on 1 trillion diverse tokens (including code and academic texts)

Model Capabilities

Text generation
Code generation
Long text processing
Instruction following

Use Cases

Content generation
Recipe generation
Generates complete cooking recipes from simple prompts
Example demonstrates ability to generate vegan banana bread recipes
Dialogue systems
Chatbot
Build dialogue systems using fine-tuned MPT-30B-Chat
Demo space available on Hugging Face
Academic research
Paper summarization
Processes lengthy academic texts and generates summaries
Benefits from training on Semantic Scholar ORC data
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase