M

Mamba 130m Hf

Developed by state-spaces
Mamba is a transformer-compatible sequence modeling model with efficient inference capabilities.
Downloads 46.83k
Release Time : 3/6/2024

Model Overview

Mamba is a sequence modeling architecture based on State Space Models (SSM), suitable for causal language modeling tasks. The model provides optimized CUDA kernel implementations, supporting efficient text generation.

Model Features

Efficient Inference
Achieves efficient inference through optimized CUDA kernels, supporting long sequence processing.
State Space Architecture
Adopts State Space Model (SSM) architecture, suitable for sequence modeling tasks.
PEFT Compatibility
Supports Parameter-Efficient Fine-Tuning (PEFT) techniques such as LoRA.

Model Capabilities

Text Generation
Language Modeling
Sequence Modeling

Use Cases

Text Generation
Dialogue Generation
Used for building chatbots or dialogue systems
Examples show the ability to generate coherent dialogue responses
Content Creation
Assists in writing and content generation
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase