đ Compumacy-Experimental-32B Quantized Model
This project provides quantized versions of the Compumacy-Experimental-32B model, which is useful for various text - generation and medical - related tasks.
đ Quick Start
If you want to start using the quantized model, you need to know how to handle GGUF files. For detailed guidance, refer to TheBloke's READMEs, which also includes instructions on concatenating multi - part files.
⨠Features
- Quantized Model: Offers weighted/imatrix quants of the base model https://huggingface.co/Daemontatox/Compumacy-Experimental-32B.
- Multiple Quant Types: Provides a variety of quant types with different sizes and qualities, sorted by size.
- Useful Resources: Includes a comparison graph of lower - quality quant types and relevant thoughts from Artefact2.
đ Documentation
About
The model is based on Daemontatox/Compumacy-Experimental-32B
and uses datasets from Daemontatox/Psy-Data-books
. The language supported is English, and it uses the transformers
library. The license is apache - 2.0
, and it is quantized by mradermacher
.
Property |
Details |
Base Model |
Daemontatox/Compumacy-Experimental-32B |
Datasets |
Daemontatox/Psy-Data-books |
Language |
English |
Library Name |
transformers |
License |
apache - 2.0 |
Quantized By |
mradermacher |
Tags |
qwen, unsloth, lora, psychology, psychiatry, clinical - assessment, mental - health, text - generation, medical |
The weighted/imatrix quants of the base model can be found at https://huggingface.co/Daemontatox/Compumacy-Experimental-32B. Static quants are available at https://huggingface.co/mradermacher/Compumacy-Experimental-32B-GGUF.
Provided Quants
The following table shows the provided quants, sorted by size (not necessarily quality. IQ - quants are often preferable over similar sized non - IQ quants):
Link |
Type |
Size/GB |
Notes |
GGUF |
i1 - IQ1_S |
7.4 |
for the desperate |
GGUF |
i1 - IQ1_M |
8.1 |
mostly desperate |
GGUF |
i1 - IQ2_XXS |
9.1 |
|
GGUF |
i1 - IQ2_XS |
10.1 |
|
GGUF |
i1 - IQ2_S |
10.6 |
|
GGUF |
i1 - IQ2_M |
11.5 |
|
GGUF |
i1 - Q2_K_S |
11.6 |
very low quality |
GGUF |
i1 - Q2_K |
12.4 |
IQ3_XXS probably better |
GGUF |
i1 - IQ3_XXS |
12.9 |
lower quality |
GGUF |
i1 - IQ3_XS |
13.8 |
|
GGUF |
i1 - Q3_K_S |
14.5 |
IQ3_XS probably better |
GGUF |
i1 - IQ3_S |
14.5 |
beats Q3_K* |
GGUF |
i1 - IQ3_M |
15.0 |
|
GGUF |
i1 - Q3_K_M |
16.1 |
IQ3_S probably better |
GGUF |
i1 - Q3_K_L |
17.4 |
IQ3_M probably better |
GGUF |
i1 - IQ4_XS |
17.8 |
|
GGUF |
i1 - Q4_0 |
18.8 |
fast, low quality |
GGUF |
i1 - Q4_K_S |
18.9 |
optimal size/speed/quality |
GGUF |
i1 - Q4_K_M |
19.9 |
fast, recommended |
GGUF |
i1 - Q4_1 |
20.7 |
|
GGUF |
i1 - Q5_K_S |
22.7 |
|
GGUF |
i1 - Q5_K_M |
23.3 |
|
GGUF |
i1 - Q6_K |
27.0 |
practically like static Q6_K |
Here is a handy graph by ikawrakow comparing some lower - quality quant types (lower is better):

And here are Artefact2's thoughts on the matter:
https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9
FAQ / Model Request
If you have any questions or want to request another model to be quantized, see https://huggingface.co/mradermacher/model_requests for answers.
đ License
The model is licensed under the apache - 2.0
license.
Thanks
I express my gratitude to my company, nethype GmbH, for allowing me to use its servers and upgrading my workstation, which enables me to do this work in my free time. Additionally, I thank @nicoboss for giving me access to his private supercomputer. This access allows me to provide many more imatrix quants at a much higher quality than I could otherwise.