L

Llama 3.2 3B Instruct Abliterated GGUF

Developed by ZeroWw
An optimized quantized model where output and embedding tensors use f16 format, while other tensors use q5_k or q6_k format, resulting in a smaller size with performance comparable to pure f16.
Downloads 20
Release Time : 10/8/2024

Model Overview

This model is a quantized version that reduces the model size while maintaining performance through optimized tensor formats. Suitable for scenarios requiring efficient inference.

Model Features

Efficient Quantization
Output and embedding tensors use f16 format, while other tensors use q5_k or q6_k format, significantly reducing model size.
Performance Retention
The quantized model's performance is comparable to pure f16 format, making it suitable for efficient inference.

Model Capabilities

Text Generation

Use Cases

Efficient Inference
Lightweight Text Generation
Suitable for text generation tasks on resource-constrained devices.
Smaller model size with performance comparable to pure f16.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase