J

Jamba V0.1 9B

Developed by TechxGenus
Jamba is a state-of-the-art hybrid SSM-Transformer architecture large language model, combining the advantages of attention mechanisms and the Mamba architecture, supporting a 256K context length, and suitable for inference on a single 80GB GPU.
Downloads 22
Release Time : 4/8/2024

Model Overview

Jamba is a pre-trained mixture-of-experts (MoE) generative text model with 12 billion active parameters and a total of 52 billion parameters across all experts. The model performs on par with or outperforms the best models of similar size in most common benchmarks.

Model Features

Hybrid Architecture
Combines the advantages of Transformer's attention mechanism and the Mamba architecture, improving model throughput.
Long Context Support
Supports context lengths of up to 256K, suitable for processing long documents and complex tasks.
Efficient Inference
Optimized implementation can handle up to 140K tokens on a single 80GB GPU, making it suitable for real-world deployment.
Mixture-of-Experts (MoE)
Adopts a mixture-of-experts architecture with 12 billion active parameters and 52 billion total parameters, balancing performance and efficiency.

Model Capabilities

Text Generation
Long Context Processing
Efficient Inference

Use Cases

Text Generation
Content Creation
Generate high-quality articles, stories, or other textual content.
Code Generation
Assist developers in generating code snippets or completing programming tasks.
Research & Development
Model Fine-Tuning
Serve as a base model for fine-tuning via the PEFT library to adapt to specific tasks.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase