B

Bitnet B1.58 2B 4T Gguf

Developed by microsoft
The first open-source, native 1-bit large language model developed by Microsoft Research, with a parameter scale of 2 billion, trained on a corpus of 4 trillion tokens.
Downloads 25.77k
Release Time : 4/15/2025

Model Overview

BitNet b1.58 2B4T is a native 1-bit large language model that demonstrates native 1-bit LLMs can achieve performance comparable to mainstream open-weight, full-precision models of similar scale, while offering significant advantages in computational efficiency (memory, energy consumption, latency).

Model Features

Native 1.58-bit quantization
Weights are quantized to ternary values {-1, 0, +1} via absolute mean quantization, while activations are quantized to 8-bit integers (per token) via absolute max quantization.
Efficient computation
Offers significant advantages in computational efficiency (memory, energy consumption, latency), with memory usage as low as 0.4GB and latency down to 29ms (CPU decoding).
Large-scale training
Trained on a corpus of 4 trillion tokens, proving that native 1-bit LLMs can achieve performance comparable to mainstream open-weight, full-precision models of similar scale.
Long-context support
Supports long-context tasks with a maximum sequence length of 4096 tokens.

Model Capabilities

Text generation
Chat dialogue
Instruction following
Mathematical reasoning
Common-sense QA

Use Cases

Dialogue systems
AI assistant
Engages in dialogue as an AI assistant to answer user questions.
Capable of generating fluent and coherent dialogue responses.
Education
Math problem solving
Solves math problems, including mathematical reasoning tasks like GSM8K and MATH-500.
Performs excellently on mathematical reasoning tasks such as GSM8K and MATH-500.
QA systems
Common-sense QA
Answers common-sense questions, such as tasks from the ARC Challenge and OpenbookQA.
Performs well on tasks like the ARC Challenge and OpenbookQA.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase