J

Jamba V0.1

Developed by ai21labs
Jamba is a state-of-the-art hybrid SSM-Transformer large language model that combines the advantages of Mamba architecture with Transformer, supporting 256K context length, surpassing models of similar scale in throughput and performance.
Downloads 6,247
Release Time : 3/28/2024

Model Overview

Jamba is the first production-grade Mamba implementation, serving as a pre-trained Mixture of Experts (MoE) text generation model with 12 billion active parameters and 52 billion total parameters. Suitable for text generation, fine-tuning, and research & development.

Model Features

Hybrid Architecture Innovation
Combines Mamba's SSM architecture with traditional Transformer to achieve throughput improvement while maintaining high performance
Ultra-Long Context Support
Supports 256K tokens context length, with a single 80GB GPU capable of processing 140K tokens
Efficient Mixture of Experts
Adopts MoE design with 52 billion total parameters but only 12 billion active parameters, balancing performance and efficiency
Production-Grade Implementation
The first practical Mamba architecture implementation for real-world applications, opening new possibilities for development

Model Capabilities

Long text generation
Knowledge Q&A
Text continuation
Instruction fine-tuning foundation

Use Cases

Research & Development
Architecture Innovation Research
Exploring the performance boundaries of hybrid SSM-Transformer architecture
Achieves or surpasses models of similar scale in multiple benchmark tests
Enterprise Applications
Long Document Processing
Utilizing 256K context length to process ultra-long documents
Maintains long-distance semantic consistency
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase