L

Llama 2 13B GPTQ

Developed by TheBloke
GPTQ quantized version of Meta's Llama 2 13B model, suitable for efficient inference
Downloads 538
Release Time : 7/18/2023

Model Overview

This is the GPTQ quantized version of Meta's Llama 2 13B model, provided by TheBloke. The model has been quantized to reduce computational resource requirements while maintaining high inference quality.

Model Features

Efficient Inference
Reduces computational resource requirements while maintaining high inference quality through GPTQ quantization technology
Multiple Quantization Options
Offers 4-bit and 8-bit quantized versions, along with different group sizes and activation order options to meet various hardware needs
ExLlama Compatibility
The 4-bit quantized version is compatible with the ExLlama inference framework

Model Capabilities

Text Generation
Dialogue Systems
Content Creation

Use Cases

Content Generation
Article Writing
Generate high-quality articles and blog content
Dialogue Systems
Build intelligent chatbots
Research & Development
NLP Research
Used for natural language processing-related research
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase