Q

Qwq 32B GGUF

Developed by tensorblock
GGUF format quantized version of QwQ-32B, suitable for local deployment and inference
Downloads 1,312
Release Time : 3/5/2025

Model Overview

This repository contains GGUF format model files for Qwen/QwQ-32B, quantized by TensorBlock's machines, compatible with llama.cpp.

Model Features

Multiple quantization versions
Provides 12 quantization versions from Q2_K to Q8_0 to meet different hardware and performance needs
llama.cpp compatibility
Compatible with llama.cpp up to commit b4823, facilitating local deployment
Chat optimization
Provides specialized prompt templates to enhance chat interaction experience

Model Capabilities

Text generation
Conversational interaction

Use Cases

Dialogue systems
Intelligent chat assistant
Deploy local chatbots
Content generation
Text creation
Generate various text content
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase