mradermacher

500 models • 91 total models in database
Sort by:

Crow-9B-Opus-4.6-Distill-Heretic_Qwen3.5-GGUF

NaNK
license:apache-2.0
182,432
16

DeepSeek-V2-Lite-GGUF

160,578
3

Lumimaid-v0.2-8B-Heretic-i1-GGUF

NaNK
license:cc-by-nc-4.0
81,830
1

Qwen3.5-27B-Claude-4.6-Opus-Reasoning-Distilled-heretic-i1-GGUF

NaNK
license:apache-2.0
79,011
3

MN-12B-Mag-Mell-R1-GGUF

static quants of https://huggingface.co/inflatebot/MN-12B-Mag-Mell-R1 weighted/imatrix quants are available at https://huggingface.co/mradermacher/MN-12B-Mag-Mell-R1-i1-GGUF Usage If you are unsure how to use GGUF files, refer to one of TheBloke's READMEs for more details, including on how to concatenate multi-part files. (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | GGUF | Q2K | 4.9 | | | GGUF | IQ3XS | 5.4 | | | GGUF | Q3KS | 5.6 | | | GGUF | IQ3S | 5.7 | beats Q3K | | GGUF | IQ3M | 5.8 | | | GGUF | Q3KM | 6.2 | lower quality | | GGUF | Q3KL | 6.7 | | | GGUF | IQ4XS | 6.9 | | | GGUF | Q4KS | 7.2 | fast, recommended | | GGUF | Q4KM | 7.6 | fast, recommended | | GGUF | Q5KS | 8.6 | | | GGUF | Q5KM | 8.8 | | | GGUF | Q6K | 10.2 | very good quality | | GGUF | Q80 | 13.1 | fast, best quality | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 See https://huggingface.co/mradermacher/modelrequests for some answers to questions you might have and/or if you want some other model quantized. I thank my company, nethype GmbH, for letting me use its servers and providing upgrades to my workstation to enable this work in my free time.

NaNK
64,426
40

GPT-OSS-Swallow-120B-RL-v0.1-i1-GGUF

NaNK
license:apache-2.0
62,357
0

OpenAI-gpt-oss-20B-INSTRUCT-Heretic-Uncensored-MXFP4-i1-GGUF

NaNK
license:apache-2.0
35,633
0

MARTIN-9B-i1-GGUF

NaNK
license:apache-2.0
34,689
1

GLM-4.7-Flash-ultra-heretic-i1-GGUF

license:mit
32,631
1

Dirty-Muse-Writer-v01-Uncensored-Erotica-NSFW-i1-GGUF

25,568
39

Apertus-70B-Instruct-2509-heretic-v2-i1-GGUF

NaNK
license:apache-2.0
24,879
0

DeepSeek-R1-Distill-Qwen-7B-Uncensored-i1-GGUF

NaNK
license:mit
24,562
62

Llama-3.1-8B-Instruct-heretic-i1-GGUF

weighted/imatrix quants of https://huggingface.co/p-e-w/Llama-3.1-8B-Instruct-heretic For a convenient overview and download list, visit our model page for this model. static quants are available at https://huggingface.co/mradermacher/Llama-3.1-8B-Instruct-heretic-GGUF Usage If you are unsure how to use GGUF files, refer to one of TheBloke's READMEs for more details, including on how to concatenate multi-part files. (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | GGUF | imatrix | 0.1 | imatrix file (for creating your own qwuants) | | GGUF | i1-IQ1S | 2.1 | for the desperate | | GGUF | i1-IQ1M | 2.3 | mostly desperate | | GGUF | i1-IQ2XXS | 2.5 | | | GGUF | i1-IQ2XS | 2.7 | | | GGUF | i1-IQ2S | 2.9 | | | GGUF | i1-IQ2M | 3.0 | | | GGUF | i1-Q2KS | 3.1 | very low quality | | GGUF | i1-Q2K | 3.3 | IQ3XXS probably better | | GGUF | i1-IQ3XXS | 3.4 | lower quality | | GGUF | i1-IQ3XS | 3.6 | | | GGUF | i1-Q3KS | 3.8 | IQ3XS probably better | | GGUF | i1-IQ3S | 3.8 | beats Q3K | | GGUF | i1-IQ3M | 3.9 | | | GGUF | i1-Q3KM | 4.1 | IQ3S probably better | | GGUF | i1-Q3KL | 4.4 | IQ3M probably better | | GGUF | i1-IQ4XS | 4.5 | | | GGUF | i1-Q40 | 4.8 | fast, low quality | | GGUF | i1-IQ4NL | 4.8 | prefer IQ4XS | | GGUF | i1-Q4KS | 4.8 | optimal size/speed/quality | | GGUF | i1-Q4KM | 5.0 | fast, recommended | | GGUF | i1-Q41 | 5.2 | | | GGUF | i1-Q5KS | 5.7 | | | GGUF | i1-Q5KM | 5.8 | | | GGUF | i1-Q6K | 6.7 | practically like static Q6K | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 See https://huggingface.co/mradermacher/modelrequests for some answers to questions you might have and/or if you want some other model quantized. I thank my company, nethype GmbH, for letting me use its servers and providing upgrades to my workstation to enable this work in my free time. Additional thanks to @nicoboss for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to.

NaNK
llama
24,110
2

Assistant_Pepe_70B-i1-GGUF

NaNK
license:llama3.1
22,683
1

Qwen3.5-122B-Turkish-Reasoning-6shard-i1-GGUF

NaNK
22,412
0

Llama3_3-Nemo-Super-Writer-49B-i1-GGUF

NaNK
llama-3
22,079
0

L3.3-MS-Nevoria-70b-heretic-i1-GGUF

NaNK
20,537
0

Qwen3.5-27B-Writer-i1-GGUF

NaNK
license:apache-2.0
20,005
3

Pokemon-Red-Qwen3-80B-i1-GGUF

NaNK
license:mit
19,998
0

MN-Violet-Lotus-12B-GGUF

NaNK
license:cc-by-4.0
19,903
37

OpenAI-gpt-oss-20B-GPT5.1-5.2-DISTILL-Heretic-Uncensored-MXFP4-i1-GGUF

NaNK
license:apache-2.0
19,711
0

Apertus-70B-Instruct-2509-heretic-v3-i1-GGUF

NaNK
license:apache-2.0
19,643
0

Deepseeker-Kunou-Qwen2.5-14b-i1-GGUF

weighted/imatrix quants of https://huggingface.co/Statuo/Deepseeker-Kunou-Qwen2.5-14b static quants are available at https://huggingface.co/mradermacher/Deepseeker-Kunou-Qwen2.5-14b-GGUF Usage If you are unsure how to use GGUF files, refer to one of TheBloke's READMEs for more details, including on how to concatenate multi-part files. (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | GGUF | i1-IQ1S | 3.7 | for the desperate | | GGUF | i1-IQ1M | 4.0 | mostly desperate | | GGUF | i1-IQ2XXS | 4.4 | | | GGUF | i1-IQ2XS | 4.8 | | | GGUF | i1-IQ2S | 5.1 | | | GGUF | i1-IQ2M | 5.5 | | | GGUF | i1-Q2KS | 5.5 | very low quality | | GGUF | i1-Q2K | 5.9 | IQ3XXS probably better | | GGUF | i1-IQ3XXS | 6.0 | lower quality | | GGUF | i1-IQ3XS | 6.5 | | | GGUF | i1-Q3KS | 6.8 | IQ3XS probably better | | GGUF | i1-IQ3S | 6.8 | beats Q3K | | GGUF | i1-IQ3M | 7.0 | | | GGUF | i1-Q3KM | 7.4 | IQ3S probably better | | GGUF | i1-Q3KL | 8.0 | IQ3M probably better | | GGUF | i1-IQ4XS | 8.2 | | | GGUF | i1-Q40 | 8.6 | fast, low quality | | GGUF | i1-IQ4NL | 8.6 | prefer IQ4XS | | GGUF | i1-Q4KS | 8.7 | optimal size/speed/quality | | GGUF | i1-Q4KM | 9.1 | fast, recommended | | GGUF | i1-Q41 | 9.5 | | | GGUF | i1-Q5KS | 10.4 | | | GGUF | i1-Q5KM | 10.6 | | | GGUF | i1-Q6K | 12.2 | practically like static Q6K | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 See https://huggingface.co/mradermacher/modelrequests for some answers to questions you might have and/or if you want some other model quantized. I thank my company, nethype GmbH, for letting me use its servers and providing upgrades to my workstation to enable this work in my free time. Additional thanks to @nicoboss for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to.

NaNK
license:apache-2.0
19,304
2

ClinAligh-30B-A3B-i1-GGUF

NaNK
license:apache-2.0
19,076
1

Trickster-Theta-4-70B-i1-GGUF

NaNK
llama-3
18,958
2

Magnum-Opus-35B-A3B-i1-GGUF

NaNK
license:apache-2.0
18,686
1

Crow-9B-Opus-4.6-Distill-Heretic_Qwen3.5-i1-GGUF

NaNK
license:apache-2.0
18,147
8

gemma-4-31b-it-heretic-ara-i1-GGUF

NaNK
license:apache-2.0
18,104
2

aum-1-70B-i1-GGUF

NaNK
llama
17,840
0

gpt-oss-120b-tainted-heresy-i1-GGUF

NaNK
license:apache-2.0
17,287
0

michaelwaves-Amoral-GPT-OSS-112E-i1-GGUF

license:apache-2.0
17,131
0

deepsex-34b-GGUF

NaNK
license:mit
16,937
4

Huihui-Qwen3-Coder-Next-abliterated-i1-GGUF

license:apache-2.0
16,806
1

Trinity-Large-TrueBase-i1-GGUF

license:apache-2.0
16,661
1

Dawn-Max-i1-GGUF

16,599
0

MedQWEN-2.5-32B-i1-GGUF

NaNK
15,902
1

Qwen3.5-27B-ultra-uncensored-heretic-v1-i1-GGUF

NaNK
license:apache-2.0
15,734
0

PE-Type-3-Nova-4B-i1-GGUF

NaNK
15,114
0

ELM-gpt-oss-20b-NSFW-v0.1-i1-GGUF

NaNK
15,093
0

Qwen3.5-27B-DS9-i1-GGUF

NaNK
license:apache-2.0
15,019
0

MiroThinker-1.7-i1-GGUF

NaNK
license:apache-2.0
14,569
0

lynx-instruct-30b-qwen3-i1-GGUF

NaNK
license:apache-2.0
14,482
0

OpenAI-gpt-oss-20B-INSTRUCT-Heretic-Uncensored-i1-GGUF

NaNK
license:apache-2.0
14,207
0

PsychAgent-Qwen3-32B-i1-GGUF

NaNK
llama-factory
14,135
2

Maenad-70B-i1-GGUF

NaNK
license:llama3
14,112
0

GLM-4.7-REAP-218B-A32B-i1-GGUF

NaNK
license:mit
14,070
1

MiniMax-M2.1-REAP-30-i1-GGUF

NaNK
license:apache-2.0
14,048
2

gemma-4-19b-a4b-it-REAP-i1-GGUF

NaNK
13,997
1

Qwen3.5-122B-A10B-abliterated-v1-i1-GGUF

NaNK
license:apache-2.0
13,632
3

MiniMax-M2-REAP-139B-A10B-i1-GGUF

NaNK
13,509
3

Void-Citrus-L3.3-70B-i1-GGUF

NaNK
13,500
2

Monika-122B-i1-GGUF

NaNK
license:cc-by-nc-4.0
13,494
1

MiniMax-M2.1-REAP-172B-A10B-i1-GGUF

NaNK
13,454
0

Rio-3.0-Open-Search-i1-GGUF

13,348
1

Samantha-big-MoE-i1-GGUF

13,224
0

SafeWork-R1-DeepSeek-70B-i1-GGUF

NaNK
license:apache-2.0
13,190
0

zen4-coder-i1-GGUF

license:apache-2.0
13,152
0

Qwen3-Next-448E-Abliterated-Instruct-i1-GGUF

license:mit
13,010
0

AReaL-tau2-merge-sft-235B-i1-GGUF

NaNK
license:apache-2.0
12,869
0

Qwen3-Coder-30B-A3B-Instruct-Heretic-i1-GGUF

NaNK
license:apache-2.0
12,856
1

Step-3.5-Flash-i1-GGUF

license:apache-2.0
12,772
0

schonsense_70B_thinkthonk-i1-GGUF

NaNK
12,745
0

Chronos-Gold-12B-1.0-i1-GGUF

NaNK
license:apache-2.0
12,724
14

Qwen2.5-32B-Instruct-heretic-i1-GGUF

NaNK
12,651
0

Qwen3.5-27B-ultra-uncensored-heretic-v2-i1-GGUF

NaNK
license:apache-2.0
12,414
0

70B_Imperious-i1-GGUF

NaNK
12,372
0

magnum-v4-12b-GGUF

NaNK
dataset:anthracite-org/c2_logs_32k_llama3_qwen2_v1.2_no_system
12,347
4

Llama-3.3-70B-Instruct-abliterated-v2-i1-GGUF

NaNK
base_model:surelio/Llama-3.3-70B-Instruct-abliterated-v1.1.1
12,327
0

BlenderCartel-llama33-70B-Pt2-i1-GGUF

NaNK
base_model:KaraKaraWitch/BlenderCartel-llama33-70B-Pt2
12,283
0

MiniMax-M2-THRIFT-55-i1-GGUF

NaNK
license:mit
12,199
2

GLM-4.6V-i1-GGUF

license:mit
12,199
1

Apertus-70B-Instruct-2509-heretic-v1-i1-GGUF

NaNK
license:apache-2.0
11,825
0

PE-Type-1-Vera-4B-i1-GGUF

NaNK
license:apache-2.0
11,774
1

Qwen3.5-35B-A3B-heretic-v2-GGUF

NaNK
license:apache-2.0
11,695
5

zen4-thinking-i1-GGUF

license:apache-2.0
11,386
0

Nanonets-OCR2-3B-GGUF

static quants of https://huggingface.co/nanonets/Nanonets-OCR2-3B For a convenient overview and download list, visit our model page for this model. weighted/imatrix quants are available at https://huggingface.co/mradermacher/Nanonets-OCR2-3B-i1-GGUF Usage If you are unsure how to use GGUF files, refer to one of TheBloke's READMEs for more details, including on how to concatenate multi-part files. (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | GGUF | mmproj-Q80 | 0.9 | multi-modal supplement | | GGUF | Q2K | 1.4 | | | GGUF | mmproj-f16 | 1.4 | multi-modal supplement | | GGUF | Q3KS | 1.6 | | | GGUF | Q3KM | 1.7 | lower quality | | GGUF | Q3KL | 1.8 | | | GGUF | IQ4XS | 1.9 | | | GGUF | Q4KS | 1.9 | fast, recommended | | GGUF | Q4KM | 2.0 | fast, recommended | | GGUF | Q5KS | 2.3 | | | GGUF | Q5KM | 2.3 | | | GGUF | Q6K | 2.6 | very good quality | | GGUF | Q80 | 3.4 | fast, best quality | | GGUF | f16 | 6.3 | 16 bpw, overkill | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 See https://huggingface.co/mradermacher/modelrequests for some answers to questions you might have and/or if you want some other model quantized. I thank my company, nethype GmbH, for letting me use its servers and providing upgrades to my workstation to enable this work in my free time.

NaNK
11,341
13

qwen-3.5-122B-uncensored-stxt-i1-GGUF

NaNK
license:apache-2.0
11,206
0

MiniMax-M2.5-CARVE-v1-BF16-i1-GGUF

NaNK
10,888
0

Ina-v11.1-i1-GGUF

NaNK
license:cc-by-nc-4.0
10,787
0

locai-l1-large-2011-i1-GGUF

license:apache-2.0
10,555
0

Cogidonia-24B-i1-GGUF

NaNK
10,517
0

Neuron-14B-i1-GGUF

NaNK
license:apache-2.0
10,227
2

Golem-70B-v1a-i1-GGUF

NaNK
10,138
0

Gemini-3-Pro-Qwen3.5-35B-A3B-i1-GGUF

NaNK
license:mit
10,044
0

zen3-nano-i1-GGUF

license:apache-2.0
10,022
0

MiniMax-M2.5-REAP-139B-A10B-i1-GGUF

NaNK
9,914
1

gpt-oss-20b-uncensored-bf16-GGUF

static quants of https://huggingface.co/huizimao/gpt-oss-20b-uncensored-bf16 For a convenient overview and download list, visit our model page for this model. weighted/imatrix quants are available at https://huggingface.co/mradermacher/gpt-oss-20b-uncensored-bf16-i1-GGUF Usage If you are unsure how to use GGUF files, refer to one of TheBloke's READMEs for more details, including on how to concatenate multi-part files. (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | GGUF | Q3KS | 12.2 | | | GGUF | Q2K | 12.2 | | | GGUF | IQ4XS | 12.3 | | | GGUF | Q3KM | 13.0 | lower quality | | GGUF | Q3KL | 13.4 | | | GGUF | Q4KS | 14.8 | fast, recommended | | GGUF | Q4KM | 15.9 | fast, recommended | | GGUF | Q5KS | 16.0 | | | GGUF | Q5KM | 17.0 | | | GGUF | Q6K | 22.3 | very good quality | | GGUF | Q80 | 22.4 | fast, best quality | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 See https://huggingface.co/mradermacher/modelrequests for some answers to questions you might have and/or if you want some other model quantized. I thank my company, nethype GmbH, for letting me use its servers and providing upgrades to my workstation to enable this work in my free time.

NaNK
license:apache-2.0
9,818
10

Qwen3.5-40B-Claude-4.6-Opus-Deckard-Heretic-Uncensored-Thinking-i1-GGUF

NaNK
license:apache-2.0
9,808
4

Kimi-Linear-48B-A3B-Instruct-i1-GGUF

NaNK
license:mit
9,460
0

Qwen2.5-VL-7B-Instruct-abliterated-GGUF

static quants of https://huggingface.co/huihui-ai/Qwen2.5-VL-7B-Instruct-abliterated For a convenient overview and download list, visit our model page for this model. weighted/imatrix quants are available at https://huggingface.co/mradermacher/Qwen2.5-VL-7B-Instruct-abliterated-i1-GGUF Usage If you are unsure how to use GGUF files, refer to one of TheBloke's READMEs for more details, including on how to concatenate multi-part files. (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | GGUF | mmproj-Q80 | 1.0 | multi-modal supplement | | GGUF | mmproj-f16 | 1.5 | multi-modal supplement | | GGUF | Q2K | 3.1 | | | GGUF | Q3KS | 3.6 | | | GGUF | Q3KM | 3.9 | lower quality | | GGUF | Q3KL | 4.2 | | | GGUF | IQ4XS | 4.4 | | | GGUF | Q4KS | 4.6 | fast, recommended | | GGUF | Q4KM | 4.8 | fast, recommended | | GGUF | Q5KS | 5.4 | | | GGUF | Q5KM | 5.5 | | | GGUF | Q6K | 6.4 | very good quality | | GGUF | Q80 | 8.2 | fast, best quality | | GGUF | f16 | 15.3 | 16 bpw, overkill | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 See https://huggingface.co/mradermacher/modelrequests for some answers to questions you might have and/or if you want some other model quantized. I thank my company, nethype GmbH, for letting me use its servers and providing upgrades to my workstation to enable this work in my free time.

NaNK
9,176
16

Gradients-Covenant-V1-i1-GGUF

NaNK
8,833
0

magnum-v4-22b-i1-GGUF

NaNK
8,826
4

trohrbaugh-Qwen3.5-122B-A10B-heretic-i1-GGUF

NaNK
license:apache-2.0
8,641
0

Qwen3-235B-A22B-abliterated-i1-GGUF

NaNK
license:apache-2.0
8,422
0

jina-reranker-v1-tiny-en-GGUF

license:apache-2.0
8,402
0

Rukun-Qwen-32B-i1-GGUF

NaNK
8,394
1

GLM-4.7-Flash-ultimate-irrefusable-heretic-i1-GGUF

license:mit
8,294
2

PE-Type-2-Alma-4B-i1-GGUF

NaNK
license:apache-2.0
8,252
2

Qwen-3.5-10.5B-Frankenmerge-Opus-4.6-Distill-i1-GGUF

NaNK
license:apache-2.0
8,244
2

MiniMax-M2.5-i1-GGUF

NaNK
8,219
1

70B_llama33_stock_unslop-i1-GGUF

NaNK
base_model:schonsense/70B_llama33_stock_unslop
8,197
0

Hypnos-i1-8B-i1-GGUF

NaNK
llama-3
8,188
1

GUI-Owl-1.5-32B-Instruct-i1-GGUF

NaNK
license:mit
7,985
0

DeepSeek-R1-Distill-Qwen-14B-Uncensored-GGUF

static quants of https://huggingface.co/nicoboss/DeepSeek-R1-Distill-Qwen-14B-Uncensored weighted/imatrix quants are available at https://huggingface.co/mradermacher/DeepSeek-R1-Distill-Qwen-14B-Uncensored-i1-GGUF Usage If you are unsure how to use GGUF files, refer to one of TheBloke's READMEs for more details, including on how to concatenate multi-part files. (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | GGUF | Q2K | 5.9 | | | GGUF | Q3KS | 6.8 | | | GGUF | Q3KM | 7.4 | lower quality | | GGUF | Q3KL | 8.0 | | | GGUF | IQ4XS | 8.3 | | | GGUF | Q4KS | 8.7 | fast, recommended | | GGUF | Q4KM | 9.1 | fast, recommended | | GGUF | Q5KS | 10.4 | | | GGUF | Q5KM | 10.6 | | | GGUF | Q6K | 12.2 | very good quality | | GGUF | Q80 | 15.8 | fast, best quality | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 See https://huggingface.co/mradermacher/modelrequests for some answers to questions you might have and/or if you want some other model quantized. I thank my company, nethype GmbH, for letting me use its servers and providing upgrades to my workstation to enable this work in my free time.

NaNK
license:mit
7,943
137

Mixtral-8x7B-Instruct-v0.1-GGUF

NaNK
license:apache-2.0
7,905
1

Qwen3.5-35B-A3B-heretic-v2-eq-v1-i1-GGUF

NaNK
license:apache-2.0
7,865
1

WeirdCompound-v1.7-24b-i1-GGUF

weighted/imatrix quants of https://huggingface.co/FlareRebellion/WeirdCompound-v1.7-24b For a convenient overview and download list, visit our model page for this model. static quants are available at https://huggingface.co/mradermacher/WeirdCompound-v1.7-24b-GGUF Usage If you are unsure how to use GGUF files, refer to one of TheBloke's READMEs for more details, including on how to concatenate multi-part files. (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | GGUF | imatrix | 0.1 | imatrix file (for creating your own qwuants) | | GGUF | i1-IQ1S | 5.4 | for the desperate | | GGUF | i1-IQ1M | 5.9 | mostly desperate | | GGUF | i1-IQ2XXS | 6.6 | | | GGUF | i1-IQ2XS | 7.3 | | | GGUF | i1-IQ2S | 7.6 | | | GGUF | i1-IQ2M | 8.2 | | | GGUF | i1-Q2KS | 8.4 | very low quality | | GGUF | i1-Q2K | 9.0 | IQ3XXS probably better | | GGUF | i1-IQ3XXS | 9.4 | lower quality | | GGUF | i1-IQ3XS | 10.0 | | | GGUF | i1-Q3KS | 10.5 | IQ3XS probably better | | GGUF | i1-IQ3S | 10.5 | beats Q3K | | GGUF | i1-IQ3M | 10.8 | | | GGUF | i1-Q3KM | 11.6 | IQ3S probably better | | GGUF | i1-Q3KL | 12.5 | IQ3M probably better | | GGUF | i1-IQ4XS | 12.9 | | | GGUF | i1-Q40 | 13.6 | fast, low quality | | GGUF | i1-Q4KS | 13.6 | optimal size/speed/quality | | GGUF | i1-Q4KM | 14.4 | fast, recommended | | GGUF | i1-Q41 | 15.0 | | | GGUF | i1-Q5KS | 16.4 | | | GGUF | i1-Q5KM | 16.9 | | | GGUF | i1-Q6K | 19.4 | practically like static Q6K | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 See https://huggingface.co/mradermacher/modelrequests for some answers to questions you might have and/or if you want some other model quantized. I thank my company, nethype GmbH, for letting me use its servers and providing upgrades to my workstation to enable this work in my free time. Additional thanks to @nicoboss for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to.

NaNK
7,847
5

OmniDimen-2-20B-Emotion-i1-GGUF

NaNK
license:mit
7,791
0

Qwen2.5-VL-7B-Abliterated-Caption-it-GGUF

static quants of https://huggingface.co/prithivMLmods/Qwen2.5-VL-7B-Abliterated-Caption-it For a convenient overview and download list, visit our model page for this model. weighted/imatrix quants are available at https://huggingface.co/mradermacher/Qwen2.5-VL-7B-Abliterated-Caption-it-i1-GGUF Usage If you are unsure how to use GGUF files, refer to one of TheBloke's READMEs for more details, including on how to concatenate multi-part files. (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | GGUF | mmproj-Q80 | 1.0 | multi-modal supplement | | GGUF | mmproj-f16 | 1.5 | multi-modal supplement | | GGUF | Q2K | 3.1 | | | GGUF | Q3KS | 3.6 | | | GGUF | Q3KM | 3.9 | lower quality | | GGUF | Q3KL | 4.2 | | | GGUF | IQ4XS | 4.4 | | | GGUF | Q4KS | 4.6 | fast, recommended | | GGUF | Q4KM | 4.8 | fast, recommended | | GGUF | Q5KS | 5.4 | | | GGUF | Q5KM | 5.5 | | | GGUF | Q6K | 6.4 | very good quality | | GGUF | Q80 | 8.2 | fast, best quality | | GGUF | f16 | 15.3 | 16 bpw, overkill | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 See https://huggingface.co/mradermacher/modelrequests for some answers to questions you might have and/or if you want some other model quantized. I thank my company, nethype GmbH, for letting me use its servers and providing upgrades to my workstation to enable this work in my free time.

NaNK
license:apache-2.0
7,784
55

Frank-27B-i1-GGUF

NaNK
license:apache-2.0
7,748
0

Strawberrylemonade-L3-70B-v1.2-heretic2-i1-GGUF

NaNK
license:llama3
7,650
0

Monika-70B-i1-GGUF

NaNK
license:apache-2.0
7,566
0

qwen35-122b-memorai-v10-sft-i1-GGUF

NaNK
7,529
0

Qwen3.5-35B-A3B-Uncensored-Aggressive-safetensors-i1-GGUF

NaNK
license:apache-2.0
7,498
1

MARTHA-9B-i1-GGUF

NaNK
license:apache-2.0
7,492
6

zen4-i1-GGUF

NaNK
license:apache-2.0
7,486
1

Qwen3.5-27B-heretic-v3-i1-GGUF

NaNK
license:apache-2.0
7,456
1

gpt2-alpaca-gpt4-GGUF

NaNK
license:mit
7,383
0

Simsema_Small-4-119B-32226-i1-GGUF

NaNK
license:apache-2.0
7,375
1

Qwen3.5-9B-heretic-i1-GGUF

NaNK
license:apache-2.0
7,291
2

gpt-oss-20b-gemini-2.5-pro-distill-i1-GGUF

NaNK
license:apache-2.0
7,222
2

Qwen2.5-Coder-14B-Abliterated-i1-GGUF

NaNK
license:apache-2.0
7,187
0

gemma-4-26B-A4B-it-heretic-ara-GGUF

NaNK
license:apache-2.0
7,168
4

Llama-70B-God-Tier-i1-GGUF

NaNK
base_model:Daga2001/Llama-70B-God-Tier
7,106
0

BereavedCompound-v1.0-24b-i1-GGUF

NaNK
7,104
2

Qwen3.5-9B-ultra-heretic-i1-GGUF

NaNK
license:apache-2.0
7,076
1

mox-tiny-1-i1-GGUF

NaNK
meta-llama
7,052
4

gemma-4-31B-it-heretic-GGUF

NaNK
license:apache-2.0
7,051
2

turkish-llm-14b-instruct-i1-GGUF

NaNK
license:apache-2.0
7,047
1

Dirty-Muse-Writer-v01-Uncensored-Erotica-NSFW-GGUF

7,023
5

Mars_27B_V.1-i1-GGUF

NaNK
6,997
3

MiniMax-M2-REAP-162B-A10B-i1-GGUF

NaNK
6,938
0

Smilodon-9B-v1-i1-GGUF

NaNK
6,870
1

Gemma3-27B-it-vl-GLM-4.7-Uncensored-Heretic-Deep-Reasoning-i1-GGUF

NaNK
license:apache-2.0
6,840
2

Poe-8B-GLM5-Opus4.6-Sonnet4.5-Kimi-Grok-Gemini-3-pro-preview-HERETIC-i1-GGUF

NaNK
license:apache-2.0
6,832
2

Qwen3.5-27B-Claude-4.6-Opus-Reasoning-Distilled-i1-GGUF

NaNK
license:apache-2.0
6,829
10

MiniMax-M2.1-REAP-50-i1-GGUF

license:apache-2.0
6,809
1

Delorme_1-OCR-7B-Post1.0-i1-GGUF

NaNK
license:apache-2.0
6,769
1

Goetia-24B-v1.1-i1-GGUF

NaNK
license:apache-2.0
6,717
0

Qwen-3.5-27B-Derestricted-i1-GGUF

NaNK
unlimited
6,649
0

Qwen3-VL-235B-A22B-Thinking-heretic-i1-GGUF

NaNK
6,639
4

Qwen3-VL-8B-Abliterated-Caption-it-i1-GGUF

weighted/imatrix quants of https://huggingface.co/prithivMLmods/Qwen3-VL-8B-Abliterated-Caption-it For a convenient overview and download list, visit our model page for this model. static quants are available at https://huggingface.co/mradermacher/Qwen3-VL-8B-Abliterated-Caption-it-GGUF This is a vision model - mmproj files (if any) will be in the static repository. Usage If you are unsure how to use GGUF files, refer to one of TheBloke's READMEs for more details, including on how to concatenate multi-part files. (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | GGUF | imatrix | 0.1 | imatrix file (for creating your own qwuants) | | GGUF | i1-IQ1S | 2.2 | for the desperate | | GGUF | i1-IQ1M | 2.4 | mostly desperate | | GGUF | i1-IQ2XXS | 2.6 | | | GGUF | i1-IQ2XS | 2.8 | | | GGUF | i1-IQ2S | 3.0 | | | GGUF | i1-IQ2M | 3.2 | | | GGUF | i1-Q2KS | 3.2 | very low quality | | GGUF | i1-Q2K | 3.4 | IQ3XXS probably better | | GGUF | i1-IQ3XXS | 3.5 | lower quality | | GGUF | i1-IQ3XS | 3.7 | | | GGUF | i1-Q3KS | 3.9 | IQ3XS probably better | | GGUF | i1-IQ3S | 3.9 | beats Q3K | | GGUF | i1-IQ3M | 4.0 | | | GGUF | i1-Q3KM | 4.2 | IQ3S probably better | | GGUF | i1-Q3KL | 4.5 | IQ3M probably better | | GGUF | i1-IQ4XS | 4.7 | | | GGUF | i1-Q40 | 4.9 | fast, low quality | | GGUF | i1-IQ4NL | 4.9 | prefer IQ4XS | | GGUF | i1-Q4KS | 4.9 | optimal size/speed/quality | | GGUF | i1-Q4KM | 5.1 | fast, recommended | | GGUF | i1-Q41 | 5.3 | | | GGUF | i1-Q5KS | 5.8 | | | GGUF | i1-Q5KM | 6.0 | | | GGUF | i1-Q6K | 6.8 | practically like static Q6K | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 See https://huggingface.co/mradermacher/modelrequests for some answers to questions you might have and/or if you want some other model quantized. I thank my company, nethype GmbH, for letting me use its servers and providing upgrades to my workstation to enable this work in my free time. Additional thanks to @nicoboss for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to.

NaNK
license:apache-2.0
6,558
3

Hathor_Sofit-L3-8B-v1-GGUF

NaNK
6,480
7

IoGPT-A1-i1-GGUF

license:apache-2.0
6,428
0

TitanForge-8B-i1-GGUF

NaNK
base_model:Stormtrooperaim/llama3.1-TitanForge-8B
6,411
0

metatune-gpt20b-R1.2-i1-GGUF

NaNK
license:apache-2.0
6,244
0

Atlas-72B-SVT-merged-i1-GGUF

NaNK
license:apache-2.0
6,197
0

Nemo-Humanities-i1-GGUF

license:apache-2.0
6,148
1

L3-8B-Stheno-v3.2-i1-GGUF

NaNK
license:cc-by-nc-4.0
6,080
16

Huihui-Tongyi-DeepResearch-30B-A3B-abliterated-i1-GGUF

NaNK
license:apache-2.0
6,065
0

Josiefied-Qwen2.5-Coder-7B-Instruct-abliterated-v1-i1-GGUF

NaNK
6,046
0

Mistral-Nemo-Batman-Venom-i1-GGUF

NaNK
license:apache-2.0
6,025
2

gemma-3-4b-it-heretic-uncensored-abliterated-Extreme-i1-GGUF

NaNK
6,020
3

HER-RM-32B-i1-GGUF

NaNK
license:apache-2.0
6,008
0

Huihui-MiroThinker-v1.0-72B-abliterated-i1-GGUF

NaNK
license:mit
6,006
1

Huihui-Qwen3-Coder-30B-A3B-Instruct-abliterated-i1-GGUF

NaNK
license:apache-2.0
5,999
16

Llama-3.3-8B-Instruct-OmniWriter-i1-GGUF

NaNK
llama
5,993
1

Llama-3-70B-Instruct-abliterated-v3-i1-GGUF

NaNK
base_model:Greytechai/Llama-3-70B-Instruct-abliterated-v3
5,971
0

Precog-24B-v1-i1-GGUF

NaNK
5,914
2

mistralai_Ministral-3-8B-Instruct-2512-abliterated-i1-GGUF

NaNK
5,888
0

Qwen3.5-27B_Homebrew-i1-GGUF

NaNK
5,848
0

Lumimaid-v0.2-70B-heretic-i1-GGUF

NaNK
license:cc-by-nc-4.0
5,827
0

Qwen3.5-9B-abliterated-i1-GGUF

NaNK
license:apache-2.0
5,783
2

MS3.2-PaintedFantasy-v3-24B-i1-GGUF

NaNK
license:apache-2.0
5,768
2

mox-tiny-1-GGUF

NaNK
meta-llama
5,757
1

KorReason-35B-Darwin-i1-GGUF

NaNK
license:apache-2.0
5,753
0

gpt-oss-4B-i1-GGUF

NaNK
5,712
1

zen4-mini-i1-GGUF

license:apache-2.0
5,661
0

Forsaken-Void-12B-i1-GGUF

NaNK
5,642
2

PyGenius1F-i1-GGUF

license:apache-2.0
5,606
1

Qwen3-42B-A3B-2507-Thinking-TOTAL-RECALL-v2-Medium-MASTER-CODER-GGUF

static quants of https://huggingface.co/DavidAU/Qwen3-42B-A3B-2507-Thinking-TOTAL-RECALL-v2-Medium-MASTER-CODER For a convenient overview and download list, visit our model page for this model. weighted/imatrix quants are available at https://huggingface.co/mradermacher/Qwen3-42B-A3B-2507-Thinking-TOTAL-RECALL-v2-Medium-MASTER-CODER-i1-GGUF Usage If you are unsure how to use GGUF files, refer to one of TheBloke's READMEs for more details, including on how to concatenate multi-part files. (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | GGUF | Q2K | 15.7 | | | GGUF | Q3KS | 18.5 | | | GGUF | Q3KM | 20.5 | lower quality | | GGUF | Q3KL | 22.1 | | | GGUF | IQ4XS | 23.0 | | | GGUF | Q4KS | 24.3 | fast, recommended | | GGUF | Q4KM | 25.8 | fast, recommended | | GGUF | Q5KS | 29.3 | | | GGUF | Q5KM | 30.2 | | | GGUF | Q6K | 34.9 | very good quality | | GGUF | Q80 | 45.2 | fast, best quality | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 See https://huggingface.co/mradermacher/modelrequests for some answers to questions you might have and/or if you want some other model quantized. I thank my company, nethype GmbH, for letting me use its servers and providing upgrades to my workstation to enable this work in my free time.

NaNK
license:apache-2.0
5,567
1

Qwen3-42B-A3B-2507-Thinking-Abliterated-uncensored-TOTAL-RECALL-v2-Medium-MASTER-CODER-i1-GGUF

weighted/imatrix quants of https://huggingface.co/DavidAU/Qwen3-42B-A3B-2507-Thinking-Abliterated-uncensored-TOTAL-RECALL-v2-Medium-MASTER-CODER For a convenient overview and download list, visit our model page for this model. static quants are available at https://huggingface.co/mradermacher/Qwen3-42B-A3B-2507-Thinking-Abliterated-uncensored-TOTAL-RECALL-v2-Medium-MASTER-CODER-GGUF Usage If you are unsure how to use GGUF files, refer to one of TheBloke's READMEs for more details, including on how to concatenate multi-part files. (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | GGUF | imatrix | 0.3 | imatrix file (for creating your own qwuants) | | GGUF | i1-IQ1S | 8.9 | for the desperate | | GGUF | i1-IQ1M | 9.8 | mostly desperate | | GGUF | i1-IQ2XXS | 11.4 | | | GGUF | i1-IQ2XS | 12.6 | | | GGUF | i1-IQ2S | 12.9 | | | GGUF | i1-IQ2M | 14.1 | | | GGUF | i1-Q2KS | 14.6 | very low quality | | GGUF | i1-Q2K | 15.7 | IQ3XXS probably better | | GGUF | i1-IQ3XXS | 16.5 | lower quality | | GGUF | i1-IQ3XS | 17.5 | | | GGUF | i1-Q3KS | 18.5 | IQ3XS probably better | | GGUF | i1-IQ3S | 18.5 | beats Q3K | | GGUF | i1-IQ3M | 18.8 | | | GGUF | i1-Q3KM | 20.5 | IQ3S probably better | | GGUF | i1-Q3KL | 22.1 | IQ3M probably better | | GGUF | i1-IQ4XS | 22.8 | | | GGUF | i1-Q40 | 24.2 | fast, low quality | | GGUF | i1-Q4KS | 24.3 | optimal size/speed/quality | | GGUF | i1-Q4KM | 25.8 | fast, recommended | | GGUF | i1-Q41 | 26.7 | | | GGUF | i1-Q5KS | 29.3 | | | GGUF | i1-Q5KM | 30.2 | | | GGUF | i1-Q6K | 34.9 | practically like static Q6K | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 See https://huggingface.co/mradermacher/modelrequests for some answers to questions you might have and/or if you want some other model quantized. I thank my company, nethype GmbH, for letting me use its servers and providing upgrades to my workstation to enable this work in my free time. Additional thanks to @nicoboss for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to.

NaNK
license:apache-2.0
5,545
9

atom-27b-i1-GGUF

NaNK
5,533
1

AutoGLM-Phone-9B-i1-GGUF

NaNK
license:mit
5,524
0

MiniMax-M2.1-REAP-40-i1-GGUF

NaNK
license:apache-2.0
5,522
0

Crow-4B-Opus-4.6-Distill-Heretic_Qwen3.5-i1-GGUF

NaNK
5,490
7

GRM2-3b-i1-GGUF

NaNK
license:apache-2.0
5,486
2

Qwen3.5-40B-Claude-4.6-Opus-Deckard-Heretic-Uncensored-Thinking-GGUF

NaNK
license:apache-2.0
5,481
5

LemonKunoichiWizardV3-GGUF

NaNK
5,481
4

Huihui-Qwen3-30B-A3B-Instruct-2507-abliterated-GGUF

static quants of https://huggingface.co/huihui-ai/Huihui-Qwen3-30B-A3B-Instruct-2507-abliterated For a convenient overview and download list, visit our model page for this model. weighted/imatrix quants are available at https://huggingface.co/mradermacher/Huihui-Qwen3-30B-A3B-Instruct-2507-abliterated-i1-GGUF Usage If you are unsure how to use GGUF files, refer to one of TheBloke's READMEs for more details, including on how to concatenate multi-part files. (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | GGUF | Q2K | 11.4 | | | GGUF | Q3KS | 13.4 | | | GGUF | Q3KM | 14.8 | lower quality | | GGUF | Q3KL | 16.0 | | | GGUF | IQ4XS | 16.7 | | | GGUF | Q4KS | 17.6 | fast, recommended | | GGUF | Q4KM | 18.7 | fast, recommended | | GGUF | Q5KS | 21.2 | | | GGUF | Q5KM | 21.8 | | | GGUF | Q6K | 25.2 | very good quality | | GGUF | Q80 | 32.6 | fast, best quality | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 See https://huggingface.co/mradermacher/modelrequests for some answers to questions you might have and/or if you want some other model quantized. I thank my company, nethype GmbH, for letting me use its servers and providing upgrades to my workstation to enable this work in my free time.

NaNK
license:apache-2.0
5,453
31

Melinoe-30B-A3B-Thinking-i1-GGUF

NaNK
license:apache-2.0
5,426
2

MarsRL-i1-GGUF

license:mit
5,422
0

zen-vl-30b-instruct-i1-GGUF

NaNK
license:apache-2.0
5,385
0

tavern-sensei-qwen3.5-35B-A3B-i1-GGUF

NaNK
license:apache-2.0
5,382
0

gpt-oss-safeguard-20b-kor-enterprise-i1-GGUF

NaNK
license:apache-2.0
5,299
0

MediumAGI-V2-i1-GGUF

NaNK
license:apache-2.0
5,289
0

Melinoe-gpt-oss-21B-A3.6B-Diluted-i1-GGUF

NaNK
5,271
0

Jade-20B-i1-GGUF

NaNK
license:apache-2.0
5,232
0

Suri-Qwen-3.5-4B-Uncensored-i1-GGUF

NaNK
5,194
1

Yanfei-v2-SamCool-i1-GGUF

license:apache-2.0
5,166
0

Huihui-Qwen3-Next-80B-A3B-Thinking-abliterated-GGUF

NaNK
license:mit
5,156
3

GPT-OSS-Swallow-20B-SFT-v0.1-i1-GGUF

NaNK
license:apache-2.0
5,148
0

Qwen3-VL-8B-Instruct-Heretic-i1-GGUF

NaNK
5,143
2

gemma-3-uncensored-i1-GGUF

license:apache-2.0
5,124
1

SynthAgent-SFT-UI-TARS-1.5-7B-i1-GGUF

NaNK
license:mit
5,119
2

Grok-3-reasoning-gemma3-12B-distilled-HF-GGUF

NaNK
license:apache-2.0
5,118
7

AuroEtherealKrix-12B-i1-GGUF

NaNK
5,110
1

gpt-oss-20b-science_full_v1-i1-GGUF

NaNK
license:apache-2.0
5,108
0

Mira-v1.12.1-27B-i1-GGUF

NaNK
5,099
0

Total04-DeepSeek-R1-Distill-Llama-70B-heretic-i1-GGUF

NaNK
base_model:CCSSNE/Total04-DeepSeek-R1-Distill-Llama-70B-heretic
5,056
0

Gemma-3-27B-Derestricted-i1-GGUF

NaNK
5,030
1

Llama-3.2-3B-Instruct-uncensored-GGUF

static quants of https://huggingface.co/chuanli11/Llama-3.2-3B-Instruct-uncensored weighted/imatrix quants are available at https://huggingface.co/mradermacher/Llama-3.2-3B-Instruct-uncensored-i1-GGUF Usage If you are unsure how to use GGUF files, refer to one of TheBloke's READMEs for more details, including on how to concatenate multi-part files. (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | GGUF | Q2K | 1.6 | | | GGUF | IQ3XS | 1.7 | | | GGUF | IQ3S | 1.8 | beats Q3K | | GGUF | Q3KS | 1.8 | | | GGUF | IQ3M | 1.9 | | | GGUF | Q3KM | 2.0 | lower quality | | GGUF | Q3KL | 2.1 | | | GGUF | IQ4XS | 2.2 | | | GGUF | Q4KS | 2.2 | fast, recommended | | GGUF | Q4KM | 2.3 | fast, recommended | | GGUF | Q5KS | 2.6 | | | GGUF | Q5KM | 2.7 | | | GGUF | Q6K | 3.1 | very good quality | | GGUF | Q80 | 3.9 | fast, best quality | | GGUF | f16 | 7.3 | 16 bpw, overkill | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 See https://huggingface.co/mradermacher/modelrequests for some answers to questions you might have and/or if you want some other model quantized. I thank my company, nethype GmbH, for letting me use its servers and providing upgrades to my workstation to enable this work in my free time.

NaNK
base_model:chuanli11/Llama-3.2-3B-Instruct-uncensored
4,997
53

Irix-12B-Model_Stock-i1-GGUF

NaNK
4,984
18

gemma-3-1b-it-heretic-extreme-uncensored-abliterated-i1-GGUF

NaNK
4,976
3

llama-joycaption-beta-one-hf-llava-GGUF

base_model:fancyfeast/llama-joycaption-beta-one-hf-llava
4,966
21

BlenderCartel-llama33-70B-Pt1-i1-GGUF

NaNK
base_model:KaraKaraWitch/BlenderCartel-llama33-70B-Pt1
4,961
0

Ministral-8B-Instruct-2410-sft-i1-GGUF

NaNK
license:apache-2.0
4,949
0

gpt-oss-20b-Derestricted-i1-GGUF

NaNK
unlimited
4,945
2

Qwen3.5-4B_Homebrew-i1-GGUF

NaNK
4,937
1

SEALION-it-Lafaek-8B-ococosda-i1-GGUF

NaNK
llama
4,928
0

GPT-OSS-Swallow-120B-RL-v0.1-GGUF

NaNK
license:apache-2.0
4,923
0

atom-80b-i1-GGUF

NaNK
license:apache-2.0
4,920
3

Qwen3.5-27B-heretic-GGUF

NaNK
license:apache-2.0
4,912
10

Huihui-GLM-4.5V-abliterated-i1-GGUF

license:mit
4,912
0

Step-3.5-Flash-REAP-149B-A11B-i1-GGUF

NaNK
4,908
2

Suri-Qwen-3.5-9B-Uncensored-i1-GGUF

NaNK
license:apache-2.0
4,891
2

Mistral-Nemo-Instruct-2407-absolute-heresy-i1-GGUF

license:apache-2.0
4,868
1

abirdus-12b-instruct-s0-i1-GGUF

NaNK
4,848
0

SVD-Qwen3-Coder-Next-Thinking-i1-GGUF

4,843
0

seed-oss-36b-chat-i1-GGUF

NaNK
4,840
0

Ken1.0-67B-i1-GGUF

NaNK
4,784
0

MARTHA-73B-Qwen2-VL-i1-GGUF

NaNK
license:apache-2.0
4,762
0

XortronCriminalComputingConfig-i1-GGUF

license:apache-2.0
4,750
8

Austral-24b-GRPO-i1-GGUF

NaNK
4,729
0

L3-SthenoMaidBlackroot-8B-V1-GGUF

NaNK
4,716
18

SP-7B-i1-GGUF

NaNK
license:mit
4,705
0

llama4-dolphin-8B-GGUF

NaNK
4,703
19

Qwen3-Next-80B-A3B-Instruct-i1-GGUF

NaNK
license:apache-2.0
4,679
1

MiMo-V2-Flash-i1-GGUF

license:mit
4,677
0

Sapphira-L3.3-70b-0.2-GGUF

static quants of https://huggingface.co/BruhzWater/Sapphira-L3.3-70b-0.2 For a convenient overview and download list, visit our model page for this model. weighted/imatrix quants are available at https://huggingface.co/mradermacher/Sapphira-L3.3-70b-0.2-i1-GGUF Usage If you are unsure how to use GGUF files, refer to one of TheBloke's READMEs for more details, including on how to concatenate multi-part files. (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | GGUF | Q2K | 26.5 | | | GGUF | Q3KS | 31.0 | | | GGUF | Q3KM | 34.4 | lower quality | | GGUF | Q3KL | 37.2 | | | GGUF | IQ4XS | 38.4 | | | GGUF | Q4KS | 40.4 | fast, recommended | | GGUF | Q4KM | 42.6 | fast, recommended | | GGUF | Q5KS | 48.8 | | | GGUF | Q5KM | 50.0 | | | PART 1 PART 2 | Q6K | 58.0 | very good quality | | PART 1 PART 2 | Q80 | 75.1 | fast, best quality | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 See https://huggingface.co/mradermacher/modelrequests for some answers to questions you might have and/or if you want some other model quantized. I thank my company, nethype GmbH, for letting me use its servers and providing upgrades to my workstation to enable this work in my free time.

NaNK
4,671
0

Luna-Qwen3.5-27B-v5-i1-GGUF

NaNK
4,649
0

glm4.1v-9b-base-sft-i1-GGUF

weighted/imatrix quants of https://huggingface.co/bountyhunterxx/glm4.1v-9b-base-sft For a convenient overview and download list, visit our model page for this model. static quants are available at https://huggingface.co/mradermacher/glm4.1v-9b-base-sft-GGUF Usage If you are unsure how to use GGUF files, refer to one of TheBloke's READMEs for more details, including on how to concatenate multi-part files. (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | GGUF | imatrix | 0.1 | imatrix file (for creating your own qwuants) | | GGUF | i1-IQ1S | 3.2 | for the desperate | | GGUF | i1-IQ1M | 3.3 | mostly desperate | | GGUF | i1-IQ2XXS | 3.5 | | | GGUF | i1-IQ2XS | 3.7 | | | GGUF | i1-IQ2S | 3.9 | | | GGUF | i1-IQ2M | 4.0 | | | GGUF | i1-Q2KS | 4.1 | very low quality | | GGUF | i1-Q2K | 4.1 | IQ3XXS probably better | | GGUF | i1-IQ3XXS | 4.3 | lower quality | | GGUF | i1-IQ3XS | 4.5 | | | GGUF | i1-Q3KS | 4.7 | IQ3XS probably better | | GGUF | i1-IQ3S | 4.7 | beats Q3K | | GGUF | i1-IQ3M | 4.8 | | | GGUF | i1-Q3KM | 5.1 | IQ3S probably better | | GGUF | i1-Q3KL | 5.3 | IQ3M probably better | | GGUF | i1-IQ4XS | 5.4 | | | GGUF | i1-IQ4NL | 5.6 | prefer IQ4XS | | GGUF | i1-Q40 | 5.6 | fast, low quality | | GGUF | i1-Q4KS | 5.9 | optimal size/speed/quality | | GGUF | i1-Q41 | 6.1 | | | GGUF | i1-Q4KM | 6.3 | fast, recommended | | GGUF | i1-Q5KS | 6.8 | | | GGUF | i1-Q5KM | 7.2 | | | GGUF | i1-Q6K | 8.4 | practically like static Q6K | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 See https://huggingface.co/mradermacher/modelrequests for some answers to questions you might have and/or if you want some other model quantized. I thank my company, nethype GmbH, for letting me use its servers and providing upgrades to my workstation to enable this work in my free time. Additional thanks to @nicoboss for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to.

NaNK
license:apache-2.0
4,631
0

TranslateGemma-4B-i1-GGUF

NaNK
llama-factory
4,624
1

Poe-8b-TOP10-Distill-Heretic-Full-i1-GGUF

NaNK
license:apache-2.0
4,621
0

INTELLECT-3V-i1-GGUF

4,609
0

Qwen3.5-9B-Casual-Thinker-i1-GGUF

NaNK
license:apache-2.0
4,603
0

Llama-3.3-70B-Instruct-heretic-i1-GGUF

NaNK
llama
4,588
0

Gemma-2-Ataraxy-v4d-9B-i1-GGUF

NaNK
4,564
3

amoral-gemma3-12B-v1-i1-GGUF

NaNK
license:apache-2.0
4,531
0

NVIDIA-Nemotron-3-Super-120B-A12B-BF16-heretic-i1-GGUF

NaNK
4,524
1

Fimbulvetr-11B-v2-GGUF

NaNK
license:cc-by-nc-4.0
4,518
18

GLM-4.7-REAP-268B-A32B-i1-GGUF

NaNK
license:mit
4,513
1

XORTRON.CriminalComputing.Q35xC46-i1-GGUF

NaNK
license:apache-2.0
4,508
0

Huihui-Qwen3-4B-Instruct-2507-abliterated-GGUF

NaNK
license:apache-2.0
4,483
5

HER-32B-i1-GGUF

NaNK
license:apache-2.0
4,478
0

Covenant72B-ChatML-bf16-i1-GGUF

NaNK
4,470
0

MiMo-V2-Flash-Base-i1-GGUF

license:mit
4,468
1

Diver-GroupRank-7B-i1-GGUF

NaNK
license:apache-2.0
4,464
0

GlotMAX-101-14B-i1-GGUF

NaNK
base_model:LLaMAX/GlotMAX-101-14B-LST
4,435
0

brayniac-Qwen3.5-27B-heretic-i1-GGUF

NaNK
license:apache-2.0
4,401
1

zen4-coder-pro-i1-GGUF

license:apache-2.0
4,396
0

AgentDoG-FG-Llama3.1-8B-i1-GGUF

NaNK
base_model:AI45Research/AgentDoG-FG-Llama3.1-8B
4,391
0

Qwen3-30B-A3B-YOYO-AutoThink-i1-GGUF

NaNK
license:apache-2.0
4,389
0

Step-3.5-Flash-REAP-121B-A11B-i1-GGUF

NaNK
4,387
0

Hulu-Med-235A22-i1-GGUF

NaNK
license:apache-2.0
4,352
0

Qwen3.5-27B-heretic-v3-GGUF

NaNK
license:apache-2.0
4,326
3

Holo2-235B-A22B-i1-GGUF

NaNK
license:cc-by-nc-4.0
4,323
0

Luna-Qwen3.5-4B-v5-i1-GGUF

NaNK
4,305
1

Arjuna-8B-i1-GGUF

NaNK
4,303
0

Olmo-3-32B-Think-i1-GGUF

NaNK
license:apache-2.0
4,258
1

nemotron-medical-tuned-70b-i1-GGUF

NaNK
4,256
0

Huihui-MiroThinker-v1.0-30B-abliterated-i1-GGUF

NaNK
license:mit
4,251
0

Qwen3-4B-Claude-Sonnet-4-Reasoning-Distill-Safetensor-GGUF

static quants of https://huggingface.co/Liontix/Qwen3-4B-Claude-Sonnet-4-Reasoning-Distill-Safetensor For a convenient overview and download list, visit our model page for this model. weighted/imatrix quants seem not to be available (by me) at this time. If they do not show up a week or so after the static ones, I have probably not planned for them. Feel free to request them by opening a Community Discussion. Usage If you are unsure how to use GGUF files, refer to one of TheBloke's READMEs for more details, including on how to concatenate multi-part files. (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | GGUF | Q2K | 1.8 | | | GGUF | Q3KS | 2.0 | | | GGUF | Q3KM | 2.2 | lower quality | | GGUF | Q3KL | 2.3 | | | GGUF | IQ4XS | 2.4 | | | GGUF | Q4KS | 2.5 | fast, recommended | | GGUF | Q4KM | 2.6 | fast, recommended | | GGUF | Q5KS | 2.9 | | | GGUF | Q5KM | 3.0 | | | GGUF | Q6K | 3.4 | very good quality | | GGUF | Q80 | 4.4 | fast, best quality | | GGUF | f16 | 8.2 | 16 bpw, overkill | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 See https://huggingface.co/mradermacher/modelrequests for some answers to questions you might have and/or if you want some other model quantized. I thank my company, nethype GmbH, for letting me use its servers and providing upgrades to my workstation to enable this work in my free time.

NaNK
4,235
11

Meta-Llama-3.1-70B-Instruct-Malaysian-i1-GGUF

NaNK
base_model:Scicom-intl/Meta-Llama-3.1-70B-Instruct-Malaysian
4,219
0

Gliese-OCR-7B-Post2.0-final-i1-GGUF

weighted/imatrix quants of https://huggingface.co/prithivMLmods/Gliese-OCR-7B-Post2.0-final For a convenient overview and download list, visit our model page for this model. static quants are available at https://huggingface.co/mradermacher/Gliese-OCR-7B-Post2.0-final-GGUF This is a vision model - mmproj files (if any) will be in the static repository. Usage If you are unsure how to use GGUF files, refer to one of TheBloke's READMEs for more details, including on how to concatenate multi-part files. (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | GGUF | imatrix | 0.1 | imatrix file (for creating your own qwuants) | | GGUF | i1-IQ1S | 2.0 | for the desperate | | GGUF | i1-IQ1M | 2.1 | mostly desperate | | GGUF | i1-IQ2XXS | 2.4 | | | GGUF | i1-IQ2XS | 2.6 | | | GGUF | i1-IQ2S | 2.7 | | | GGUF | i1-IQ2M | 2.9 | | | GGUF | i1-Q2KS | 2.9 | very low quality | | GGUF | i1-Q2K | 3.1 | IQ3XXS probably better | | GGUF | i1-IQ3XXS | 3.2 | lower quality | | GGUF | i1-IQ3XS | 3.4 | | | GGUF | i1-Q3KS | 3.6 | IQ3XS probably better | | GGUF | i1-IQ3S | 3.6 | beats Q3K | | GGUF | i1-IQ3M | 3.7 | | | GGUF | i1-Q3KM | 3.9 | IQ3S probably better | | GGUF | i1-Q3KL | 4.2 | IQ3M probably better | | GGUF | i1-IQ4XS | 4.3 | | | GGUF | i1-IQ4NL | 4.5 | prefer IQ4XS | | GGUF | i1-Q40 | 4.5 | fast, low quality | | GGUF | i1-Q4KS | 4.6 | optimal size/speed/quality | | GGUF | i1-Q4KM | 4.8 | fast, recommended | | GGUF | i1-Q41 | 5.0 | | | GGUF | i1-Q5KS | 5.4 | | | GGUF | i1-Q5KM | 5.5 | | | GGUF | i1-Q6K | 6.4 | practically like static Q6K | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 See https://huggingface.co/mradermacher/modelrequests for some answers to questions you might have and/or if you want some other model quantized. I thank my company, nethype GmbH, for letting me use its servers and providing upgrades to my workstation to enable this work in my free time. Additional thanks to @nicoboss for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to.

NaNK
license:apache-2.0
4,200
1

Famino-12B-Model_Stock-i1-GGUF

NaNK
license:apache-2.0
4,197
1

HERETICSEEK-7B-Ditill-i1-GGUF

NaNK
license:apache-2.0
4,191
1

Skyfall-31B-v4.1-heretic2-i1-GGUF

NaNK
4,175
0

Broken-Tutu-24B-Unslop-v2.0-GGUF

NaNK
license:apache-2.0
4,116
3

Qwen2.5-VL-7B-NSFW-Caption-V4-GGUF

static quants of https://huggingface.co/thesby/Qwen2.5-VL-7B-NSFW-Caption-V4 For a convenient overview and download list, visit our model page for this model. weighted/imatrix quants are available at https://huggingface.co/mradermacher/Qwen2.5-VL-7B-NSFW-Caption-V4-i1-GGUF Usage If you are unsure how to use GGUF files, refer to one of TheBloke's READMEs for more details, including on how to concatenate multi-part files. (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | GGUF | mmproj-Q80 | 0.8 | multi-modal supplement | | GGUF | mmproj-f16 | 1.5 | multi-modal supplement | | GGUF | Q2K | 3.2 | | | GGUF | Q3KS | 3.6 | | | GGUF | Q3KM | 4.0 | lower quality | | GGUF | Q3KL | 4.2 | | | GGUF | IQ4XS | 4.4 | | | GGUF | Q4KS | 4.6 | fast, recommended | | GGUF | Q4KM | 4.8 | fast, recommended | | GGUF | Q5KS | 5.4 | | | GGUF | Q5KM | 5.5 | | | GGUF | Q6K | 6.4 | very good quality | | GGUF | Q80 | 8.2 | fast, best quality | | GGUF | f16 | 15.4 | 16 bpw, overkill | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 See https://huggingface.co/mradermacher/modelrequests for some answers to questions you might have and/or if you want some other model quantized. I thank my company, nethype GmbH, for letting me use its servers and providing upgrades to my workstation to enable this work in my free time.

NaNK
license:apache-2.0
4,114
11

Chaos-Unknown-12b-i1-GGUF

NaNK
4,091
1

70B_neolithic_rabbit-i1-GGUF

NaNK
4,087
0

atom-80b-GGUF

NaNK
license:apache-2.0
4,086
2

The_Creeping_Darkness-X2-16B-i1-GGUF

NaNK
license:apache-2.0
4,050
3

Llama-3.2-3B-Instruct-heretic-ablitered-uncensored-i1-GGUF

NaNK
base_model:DavidAU/Llama-3.2-3B-Instruct-heretic-ablitered-uncensored
4,050
1

Llama-3-Swallow-8B-Instruct-v0.1-kokoroe-i1-GGUF

NaNK
base_model:retrieva-jp/Llama-3-Swallow-8B-Instruct-v0.1-kokoroe
4,035
0

Qwen3-Next-80B-A3B-Thinking-i1-GGUF

NaNK
license:apache-2.0
4,034
0

70B_Triage-i1-GGUF

NaNK
4,022
0

Qwen2.5-Coder-7B-Abliterated-i1-GGUF

NaNK
license:apache-2.0
3,985
1

Qwen3-VL-REAP-145B-A22B-i1-GGUF

NaNK
license:apache-2.0
3,984
0

Harmonic-27B-i1-GGUF

NaNK
license:apache-2.0
3,983
0

Qwen3.5-Antirep-27B-i1-GGUF

NaNK
license:apache-2.0
3,983
0

Qwen3-Next-416E-Abliterated-Instruct-i1-GGUF

license:mit
3,974
0

MiniMax-M2.1-REAP-40-GGUF

NaNK
license:apache-2.0
3,963
12

SEX_ROLEPLAY-3.2-1B-i1-GGUF

NaNK
license:apache-2.0
3,957
7

Broken-Tutu-24B-Transgression-v2.0-i1-GGUF

NaNK
license:apache-2.0
3,955
3

Qwen3-VL-Reranker-8B-GGUF

NaNK
license:apache-2.0
3,950
3

MiniMax-M2.1-REAP-139B-A10B-i1-GGUF

NaNK
3,900
2

Scarlet-Seraph-12B-i1-GGUF

NaNK
3,889
1

Qwen3-Next-80B-A3B-Thinking-GRPO-Uncensored-i1-GGUF

NaNK
license:apache-2.0
3,885
1

Qwen3-0.6B-Qrazy-Qoder-i1-GGUF

NaNK
3,865
0

Suri-Qwen-3.5-4B-Uncensored-Low-i1-GGUF

NaNK
3,853
0

aidc-llm-laos-12b-i1-GGUF

NaNK
3,839
0

NVIDIA-Nemotron-3-Super-120B-A12B-BF16-i1-GGUF

NaNK
3,822
0

14B-Qwen2.5-Kunou-v1-GGUF

NaNK
3,817
3

Solar-Open-100B-i1-GGUF

NaNK
3,806
6

Olmo-3-7B-RLZero-Mix-i1-GGUF

NaNK
license:apache-2.0
3,806
0

Llama3.2-24B-A3B-II-Dark-Champion-INSTRUCT-Heretic-Abliterated-Uncensored-i1-GGUF

NaNK
Llama 3.2 MOE
3,778
0

Qwen2.5-VL-7B-Instruct-GGUF

NaNK
3,776
3

Emerald-Wyvern-12B-i1-GGUF

NaNK
3,761
2

Olmo-3-7B-Think-i1-GGUF

NaNK
license:apache-2.0
3,745
1

gemma-3-27b-it-heretic-v2-i1-GGUF

NaNK
3,744
2

Qwen3-VL-8B-Instruct-abliterated-v2.0-i1-GGUF

NaNK
license:apache-2.0
3,744
1

Qwen3.5-4B-heretic-GGUF

NaNK
license:apache-2.0
3,739
3

OpenAI-gpt-oss-20B-INSTRUCT-Heretic-Uncensored-MXFP4-GGUF

NaNK
license:apache-2.0
3,731
0

Qwen3.5-27B-heretic-v2-i1-GGUF

NaNK
license:apache-2.0
3,719
6

DECS_7B-i1-GGUF

NaNK
3,714
2

Step-3.5-Flash-REAP-128B-A11B-i1-GGUF

NaNK
license:apache-2.0
3,709
0

Magistry-24B-v1.1-i1-GGUF

NaNK
license:apache-2.0
3,703
1

wraith-8b-i1-GGUF

weighted/imatrix quants of https://huggingface.co/vanta-research/wraith-8b For a convenient overview and download list, visit our model page for this model. static quants are available at https://huggingface.co/mradermacher/wraith-8b-GGUF Usage If you are unsure how to use GGUF files, refer to one of TheBloke's READMEs for more details, including on how to concatenate multi-part files. (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | GGUF | imatrix | 0.1 | imatrix file (for creating your own qwuants) | | GGUF | i1-IQ1S | 2.1 | for the desperate | | GGUF | i1-IQ1M | 2.3 | mostly desperate | | GGUF | i1-IQ2XXS | 2.5 | | | GGUF | i1-IQ2XS | 2.7 | | | GGUF | i1-IQ2S | 2.9 | | | GGUF | i1-IQ2M | 3.0 | | | GGUF | i1-Q2KS | 3.1 | very low quality | | GGUF | i1-Q2K | 3.3 | IQ3XXS probably better | | GGUF | i1-IQ3XXS | 3.4 | lower quality | | GGUF | i1-IQ3XS | 3.6 | | | GGUF | i1-Q3KS | 3.8 | IQ3XS probably better | | GGUF | i1-IQ3S | 3.8 | beats Q3K | | GGUF | i1-IQ3M | 3.9 | | | GGUF | i1-Q3KM | 4.1 | IQ3S probably better | | GGUF | i1-Q3KL | 4.4 | IQ3M probably better | | GGUF | i1-IQ4XS | 4.5 | | | GGUF | i1-Q40 | 4.8 | fast, low quality | | GGUF | i1-IQ4NL | 4.8 | prefer IQ4XS | | GGUF | i1-Q4KS | 4.8 | optimal size/speed/quality | | GGUF | i1-Q4KM | 5.0 | fast, recommended | | GGUF | i1-Q41 | 5.2 | | | GGUF | i1-Q5KS | 5.7 | | | GGUF | i1-Q5KM | 5.8 | | | GGUF | i1-Q6K | 6.7 | practically like static Q6K | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 See https://huggingface.co/mradermacher/modelrequests for some answers to questions you might have and/or if you want some other model quantized. I thank my company, nethype GmbH, for letting me use its servers and providing upgrades to my workstation to enable this work in my free time. Additional thanks to @nicoboss for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to.

NaNK
llama
3,693
3

KQ_Omni-12B-v1-i1-GGUF

NaNK
license:apache-2.0
3,682
1

Deepseek-R1-Distill-NSFW-RPv1-GGUF

NaNK
3,680
48

ASTRA-14B-Thinking-v1-i1-GGUF

NaNK
license:apache-2.0
3,677
0

Poe-8B-GLM5-Opus4.6-Sonnet4.5-Kimi-Grok-Gemini-3-pro-preview-HERETIC-GGUF

NaNK
license:apache-2.0
3,661
3

OpenELM-3B-Instruct-GGUF

NaNK
3,660
0

Qwen_Uncensored-i1-GGUF

license:mit
3,656
0

hito-1.7b-i1-GGUF

NaNK
llama-cpp
3,618
0

survey-bot-qwen3-vl-32b-i1-GGUF

NaNK
license:apache-2.0
3,606
1

ASID-Captioner-7B-i1-GGUF

NaNK
license:apache-2.0
3,606
0

Qwen3-Coder-Next-Base-i1-GGUF

license:apache-2.0
3,602
0

MathSmith-DS-Qwen-7B-LongCoT-i1-GGUF

NaNK
llama-factory
3,589
1

DeepSeek-V2-Lite-Chat-Uncensored-Unbiased-Reasoner-GGUF

license:llama3.3
3,571
12

Stellar-Umbra-12B-i1-GGUF

NaNK
3,567
1

Hulu-Med-30A3-i1-GGUF

NaNK
license:apache-2.0
3,559
0

L3.2-3B-Herthea-i1-GGUF

NaNK
llama-3
3,551
0

GUI-Owl-1.5-8B-Think-i1-GGUF

NaNK
license:mit
3,544
0

Trinity-Mini-Base-i1-GGUF

license:apache-2.0
3,534
0

GLM-4.7-Flash-REAP-23B-A3B-absolute-heresy-i1-GGUF

NaNK
license:mit
3,532
0

Llama-3.1-EstLLM-8B-0525-i1-GGUF

NaNK
base_model:tartuNLP/Llama-3.1-EstLLM-8B-0525
3,532
0

Tankie-DPE-12b-SFT-i1-GGUF

NaNK
license:apache-2.0
3,526
0

olmo-v2-stage3-lexifreak-heretic-v1-i1-GGUF

NaNK
3,525
0

gemma3-27b-abliterated-dpo-i1-GGUF

weighted/imatrix quants of https://huggingface.co/summykai/gemma3-27b-abliterated-dpo For a convenient overview and download list, visit our model page for this model. static quants are available at https://huggingface.co/mradermacher/gemma3-27b-abliterated-dpo-GGUF This is a vision model - mmproj files (if any) will be in the static repository. Usage If you are unsure how to use GGUF files, refer to one of TheBloke's READMEs for more details, including on how to concatenate multi-part files. (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | GGUF | i1-IQ1S | 6.4 | for the desperate | | GGUF | i1-IQ1M | 6.9 | mostly desperate | | GGUF | i1-IQ2XXS | 7.8 | | | GGUF | i1-IQ2XS | 8.5 | | | GGUF | i1-IQ2S | 8.9 | | | GGUF | i1-IQ2M | 9.6 | | | GGUF | i1-Q2KS | 9.9 | very low quality | | GGUF | i1-Q2K | 10.6 | IQ3XXS probably better | | GGUF | i1-IQ3XXS | 10.8 | lower quality | | GGUF | i1-IQ3XS | 11.7 | | | GGUF | i1-IQ3S | 12.3 | beats Q3K | | GGUF | i1-Q3KS | 12.3 | IQ3XS probably better | | GGUF | i1-IQ3M | 12.6 | | | GGUF | i1-Q3KM | 13.5 | IQ3S probably better | | GGUF | i1-Q3KL | 14.6 | IQ3M probably better | | GGUF | i1-IQ4XS | 14.9 | | | GGUF | i1-Q40 | 15.7 | fast, low quality | | GGUF | i1-Q4KS | 15.8 | optimal size/speed/quality | | GGUF | i1-Q4KM | 16.6 | fast, recommended | | GGUF | i1-Q41 | 17.3 | | | GGUF | i1-Q5KS | 18.9 | | | GGUF | i1-Q5KM | 19.4 | | | GGUF | i1-Q6K | 22.3 | practically like static Q6K | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 See https://huggingface.co/mradermacher/modelrequests for some answers to questions you might have and/or if you want some other model quantized. I thank my company, nethype GmbH, for letting me use its servers and providing upgrades to my workstation to enable this work in my free time. Additional thanks to @nicoboss for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to.

NaNK
3,518
13

Mira-v1.16-Ties-27B-i1-GGUF

NaNK
3,509
0

Chimera-DeepSeek-NSFW-8B-GGUF

NaNK
base_model:ForSureTesterSim/DeepLlama-Chimera-NSFW-8B
3,496
5

Hermes-4-70B-heretic-i1-GGUF

NaNK
Llama-3.1
3,489
0

Nemo-2407-Based-Instruct-DeLERP-0.7-12B-i1-GGUF

NaNK
license:apache-2.0
3,483
1

L3.1-Apluv3-8B-i1-GGUF

NaNK
3,476
1

proxima-ocr-d.markdown-post3.0.l-i1-GGUF

license:apache-2.0
3,444
0

apertus-12b-healed-s0-i1-GGUF

NaNK
3,442
0

NemoMix-Unleashed-12B-i1-GGUF

NaNK
license:apache-2.0
3,422
19

Olmo-3-32B-Think-SFT-i1-GGUF

NaNK
license:apache-2.0
3,421
0

Llama3.2-30B-A3B-II-Dark-Champion-INSTRUCT-Heretic-Abliterated-Uncensored-i1-GGUF

NaNK
Llama 3.2 MOE
3,419
0

DeepSeek-R1-Distill-Llama-8B-Abliterated-i1-GGUF

NaNK
base_model:stepenZEN/DeepSeek-R1-Distill-Llama-8B-Abliterated
3,416
6

RP-king-12b-i1-GGUF

NaNK
license:apache-2.0
3,413
1

Qwen3.5-35B-A3B-Claude-4.6-Opus-Reasoning-Distilled-i1-GGUF

NaNK
license:apache-2.0
3,409
0

Qwopus-MoE-35B-A3B-i1-GGUF

NaNK
license:apache-2.0
3,405
1

DR-Tulu-8B-i1-GGUF

NaNK
license:apache-2.0
3,397
0

mox-small-1-i1-GGUF

NaNK
license:apache-2.0
3,389
0

maya1-i1-GGUF

NaNK
license:apache-2.0
3,377
2

PE-Type-1-Vera-4B-GGUF

NaNK
license:apache-2.0
3,368
3

mini-magnum-12b-v1.1-GGUF

NaNK
license:apache-2.0
3,358
12

Sunlit-Shadow-12B-i1-GGUF

NaNK
3,353
1

Llama3.1-DeepDilemma-V1-8B-i1-GGUF

weighted/imatrix quants of https://huggingface.co/Yuma42/Llama3.1-DeepDilemma-V1-8B For a convenient overview and download list, visit our model page for this model. static quants are available at https://huggingface.co/mradermacher/Llama3.1-DeepDilemma-V1-8B-GGUF Usage If you are unsure how to use GGUF files, refer to one of TheBloke's READMEs for more details, including on how to concatenate multi-part files. (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | GGUF | imatrix | 0.1 | imatrix file (for creating your own qwuants) | | GGUF | i1-IQ1S | 2.1 | for the desperate | | GGUF | i1-IQ1M | 2.3 | mostly desperate | | GGUF | i1-IQ2XXS | 2.5 | | | GGUF | i1-IQ2XS | 2.7 | | | GGUF | i1-IQ2S | 2.9 | | | GGUF | i1-IQ2M | 3.0 | | | GGUF | i1-Q2KS | 3.1 | very low quality | | GGUF | i1-Q2K | 3.3 | IQ3XXS probably better | | GGUF | i1-IQ3XXS | 3.4 | lower quality | | GGUF | i1-IQ3XS | 3.6 | | | GGUF | i1-Q3KS | 3.8 | IQ3XS probably better | | GGUF | i1-IQ3S | 3.8 | beats Q3K | | GGUF | i1-IQ3M | 3.9 | | | GGUF | i1-Q3KM | 4.1 | IQ3S probably better | | GGUF | i1-Q3KL | 4.4 | IQ3M probably better | | GGUF | i1-IQ4XS | 4.5 | | | GGUF | i1-Q40 | 4.8 | fast, low quality | | GGUF | i1-IQ4NL | 4.8 | prefer IQ4XS | | GGUF | i1-Q4KS | 4.8 | optimal size/speed/quality | | GGUF | i1-Q4KM | 5.0 | fast, recommended | | GGUF | i1-Q41 | 5.2 | | | GGUF | i1-Q5KS | 5.7 | | | GGUF | i1-Q5KM | 5.8 | | | GGUF | i1-Q6K | 6.7 | practically like static Q6K | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 See https://huggingface.co/mradermacher/modelrequests for some answers to questions you might have and/or if you want some other model quantized. I thank my company, nethype GmbH, for letting me use its servers and providing upgrades to my workstation to enable this work in my free time. Additional thanks to @nicoboss for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to.

NaNK
Yuma42/Llama3.1-StableRoots-V0.5-8B
3,349
3

DeepSeek-R1-Distill-Qwen-14B-abliterated-i1-GGUF

NaNK
3,346
18

ARC-Base-8B-i1-GGUF

NaNK
llama
3,345
1

Qwen3.5-4B-Claude-Opus-Reasoning-i1-GGUF

NaNK
license:apache-2.0
3,341
2

MN-CaptainErisNebula-12B-Chimera-v1.1-heretic-uncensored-abliterated-i1-GGUF

NaNK
3,340
2

Lokis_Veil-8B-i1-GGUF

NaNK
llama-3
3,338
1

VITAL-7B-i1-GGUF

NaNK
license:apache-2.0
3,323
0

littlemonster-reasoning-12B-QKVO-heretic-HF-i1-GGUF

NaNK
license:apache-2.0
3,316
0

OctoThinker-8B-Long-Base-i1-GGUF

NaNK
license:llama3.2
3,312
0

Monika-12B-i1-GGUF

NaNK
license:apache-2.0
3,306
0

EtherealKrix-12B-i1-GGUF

NaNK
3,299
0

Llama3.1-70B-Chinese-Chat-GGUF

NaNK
llama-factory
3,296
1

Qwen3-30B-A3B-abliterated-erotic-i1-GGUF

NaNK
license:apache-2.0
3,295
17

OLMo-2-1124-13B-Instruct-32k-Context-ChatML-i1-GGUF

NaNK
3,295
0

Cydonia-v4.1-MS3.2-Magnum-Diamond-24B-i1-GGUF

NaNK
license:apache-2.0
3,289
2

The_Croupier-3.2-1B-i1-GGUF

NaNK
license:llama3.2
3,287
0

ALIA-40b-instruct-2601-i1-GGUF

NaNK
license:apache-2.0
3,281
0

DynamicRAG-8B-i1-GGUF

NaNK
license:apache-2.0
3,270
1

Gemma-3-4B-THINKING-i1-GGUF

NaNK
3,254
1

Huihui-Qwen3-Coder-30B-A3B-Instruct-abliterated-GGUF

static quants of https://huggingface.co/huihui-ai/Huihui-Qwen3-Coder-30B-A3B-Instruct-abliterated For a convenient overview and download list, visit our model page for this model. weighted/imatrix quants are available at https://huggingface.co/mradermacher/Huihui-Qwen3-Coder-30B-A3B-Instruct-abliterated-i1-GGUF Usage If you are unsure how to use GGUF files, refer to one of TheBloke's READMEs for more details, including on how to concatenate multi-part files. (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | GGUF | Q2K | 11.4 | | | GGUF | Q3KS | 13.4 | | | GGUF | Q3KM | 14.8 | lower quality | | GGUF | Q3KL | 16.0 | | | GGUF | IQ4XS | 16.7 | | | GGUF | Q4KS | 17.6 | fast, recommended | | GGUF | Q4KM | 18.7 | fast, recommended | | GGUF | Q5KS | 21.2 | | | GGUF | Q5KM | 21.8 | | | GGUF | Q6K | 25.2 | very good quality | | GGUF | Q80 | 32.6 | fast, best quality | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 See https://huggingface.co/mradermacher/modelrequests for some answers to questions you might have and/or if you want some other model quantized. I thank my company, nethype GmbH, for letting me use its servers and providing upgrades to my workstation to enable this work in my free time.

NaNK
license:apache-2.0
3,249
3

VibeThinker-1.5B-i1-GGUF

NaNK
license:mit
3,230
1

Ministral-3-3B-Base-2512-i1-GGUF

NaNK
license:apache-2.0
3,230
0

Monika-24B-i1-GGUF

NaNK
license:apache-2.0
3,230
0

GeneralChat-Llama3.2-3B-DPO-i1-GGUF

NaNK
llama
3,228
0

nova-jais-2-70b-v2-i1-GGUF

NaNK
3,207
0

MiniMax-M2-THRIFT-i1-GGUF

weighted/imatrix quants of https://huggingface.co/VibeStudio/MiniMax-M2-THRIFT For a convenient overview and download list, visit our model page for this model. static quants are available at https://huggingface.co/mradermacher/MiniMax-M2-THRIFT-GGUF Usage If you are unsure how to use GGUF files, refer to one of TheBloke's READMEs for more details, including on how to concatenate multi-part files. (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | GGUF | imatrix | 0.5 | imatrix file (for creating your own qwuants) | | GGUF | i1-IQ1S | 35.3 | for the desperate | | GGUF | i1-IQ1M | 39.1 | mostly desperate | | GGUF | i1-IQ2XXS | 45.5 | | | PART 1 PART 2 | i1-IQ2XS | 50.7 | | | PART 1 PART 2 | i1-IQ2S | 51.6 | | | PART 1 PART 2 | i1-IQ2M | 56.7 | | | PART 1 PART 2 | i1-Q2KS | 58.7 | very low quality | | PART 1 PART 2 | i1-Q2K | 63.0 | IQ3XXS probably better | | PART 1 PART 2 | i1-IQ3XXS | 66.5 | lower quality | | PART 1 PART 2 | i1-IQ3XS | 70.6 | | | PART 1 PART 2 | i1-Q3KS | 74.6 | IQ3XS probably better | | PART 1 PART 2 | i1-IQ3S | 74.6 | beats Q3K | | PART 1 PART 2 | i1-IQ3M | 75.6 | | | PART 1 PART 2 | i1-Q3KM | 82.6 | IQ3S probably better | | PART 1 PART 2 | i1-Q3KL | 89.4 | IQ3M probably better | | PART 1 PART 2 | i1-IQ4XS | 92.1 | | | PART 1 PART 2 | i1-Q40 | 97.8 | fast, low quality | | PART 1 PART 2 | i1-Q4KS | 98.2 | optimal size/speed/quality | | PART 1 PART 2 PART 3 | i1-Q4KM | 104.5 | fast, recommended | | PART 1 PART 2 PART 3 | i1-Q41 | 108.2 | | | PART 1 PART 2 PART 3 | i1-Q5KS | 118.9 | | | PART 1 PART 2 PART 3 | i1-Q5KM | 122.5 | | | PART 1 PART 2 PART 3 | i1-Q6K | 141.7 | practically like static Q6K | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 See https://huggingface.co/mradermacher/modelrequests for some answers to questions you might have and/or if you want some other model quantized. I thank my company, nethype GmbH, for letting me use its servers and providing upgrades to my workstation to enable this work in my free time. Additional thanks to @nicoboss for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to.

license:mit
3,192
11

Qwen3.5-9B-ultra-heretic-GGUF

NaNK
license:apache-2.0
3,178
4

Big-Tiger-Gemma-27B-v3-heretic-i1-GGUF

NaNK
3,165
0

Qwen3.5-27B-ultimate-heretic-i1-GGUF

NaNK
license:apache-2.0
3,163
0

Ministral-3-8B-Reasoning-2512-i1-GGUF

NaNK
license:apache-2.0
3,156
0

HereticAggressive-CoT-i1-GGUF

license:apache-2.0
3,152
1

aquif-Spatial-7B-i1-GGUF

NaNK
license:apache-2.0
3,152
1

abirdus-12b-instruct-i1-GGUF

NaNK
3,130
0

Qwen3-32B-Uncensored-GGUF

NaNK
license:apache-2.0
3,124
16

magnum-v4-9b-abliterated-i1-GGUF

NaNK
dataset:anthracite-org/c2_logs_16k_llama_v1.1
3,122
1

Esperpento-1B-i1-GGUF

NaNK
3,112
0

Qwen3.5-9B-Claude-4.6-OS-INSTRUCT-i1-GGUF

NaNK
license:apache-2.0
3,098
0

GLM-4.5-Architect-106B-A12B-i1-GGUF

NaNK
license:mit
3,086
0

mox-small-1-GGUF

NaNK
license:apache-2.0
3,077
1

Nomi-1.0-3b-i1-GGUF

NaNK
llama
3,077
0

Precog-24B-v1-heretic-i1-GGUF

NaNK
3,074
0

Llama3.1-DeluXeOne-8B-i1-GGUF

NaNK
base_model:Yuma42/Llama3.1-DeluXeOne-8B
3,072
0

Seed-OSS-36B-Instruct-MPOA-v1-i1-GGUF

NaNK
license:apache-2.0
3,071
0

Olmo-3-7B-Instruct-DPO-i1-GGUF

NaNK
license:apache-2.0
3,067
0

Jackdaw-30B-A3B-i1-GGUF

NaNK
3,058
0

WeirdDolphinPersonalityMechanism-Mistral-24B-i1-GGUF

NaNK
license:apache-2.0
3,044
0

CoPaw-Flash-9B-GGUF

NaNK
license:apache-2.0
3,043
2

Dans-PersonalityEngine-V1.3.0-24b-i1-GGUF

NaNK
license:apache-2.0
3,042
3

Qwen3-30B-A3B-Thinking-2507-Gemini-2.5-Flash-Distill-i1-GGUF

NaNK
license:apache-2.0
3,039
0

Special-Virus-3.2-1B-i1-GGUF

NaNK
license:llama3.2
3,033
0

TildeOpen-30b-ENLV-ChatML-instruct-i1-GGUF

NaNK
3,027
0

Fara-7B-i1-GGUF

NaNK
license:mit
3,021
1

Mistral-Small-3_2-24B-Instruct-2506-antislop.v2-i1-GGUF

NaNK
3,019
0

JOSIE-4B-Thinking-i1-GGUF

NaNK
license:mit
3,008
2

GlotMAX-101-8B-i1-GGUF

NaNK
base_model:LLaMAX/GlotMAX-101-8B-LST
3,008
1

Qwen3-VL-8B-Interleave-Thinking-i1-GGUF

NaNK
license:apache-2.0
3,008
0

GigaChat-20B-A3B-instruct-bf16-i1-GGUF

NaNK
license:mit
3,001
0

Qwen2.5-VL-7B-V1-i1-GGUF

NaNK
license:apache-2.0
2,994
1

Bakti-8B-Base-i1-GGUF

NaNK
license:apache-2.0
2,987
0

Qwen2.5-32B-Cyberpunk-Storyteller-v2-i1-GGUF

NaNK
license:apache-2.0
2,971
0

Olmo-3-7B-Instruct-SFT-i1-GGUF

NaNK
license:apache-2.0
2,969
0

Llama-3.3-70B-Instruct-abliterated-v2-GGUF

NaNK
base_model:surelio/Llama-3.3-70B-Instruct-abliterated-v1.1.1
2,959
0

Huihui-Qwen3-Next-80B-A3B-Instruct-abliterated-GGUF

NaNK
license:mit
2,958
0

ALIA-40b-i1-GGUF

NaNK
license:apache-2.0
2,953
4

Ahma-2-4B-Instruct-i1-GGUF

NaNK
license:apache-2.0
2,951
1

DeepSeek-V3.1-Nex-N1.1-i1-GGUF

NaNK
license:apache-2.0
2,947
0

Qwen2.5-7B-Kids-SciFi-i1-GGUF

NaNK
license:apache-2.0
2,946
1

The_Darkside-16.6B-i1-GGUF

NaNK
license:apache-2.0
2,928
2

Broken-Tutu-24B-Unslop-v2.0-i1-GGUF

NaNK
license:apache-2.0
2,927
5

sundae-v716-generate-direct-4b-i1-GGUF

NaNK
license:cc-by-nc-4.0
2,902
0

SAI-DeepCoder-14B-Preview-unsloth-v1.0-i1-GGUF

NaNK
license:mit
2,890
0

Cicikus_v2_3B-i1-GGUF

NaNK
llama-3.2
2,889
0

PG67A-W-Serum.Test-3.2-1B-i1-GGUF

NaNK
license:llama3.2
2,886
0

Gemma-4-31B-Cognitive-Unshackled-GGUF

NaNK
license:apache-2.0
2,882
1

Qwen3-VL-32B-Instruct-abliterated-v1-i1-GGUF

NaNK
license:apache-2.0
2,870
0

Unbound-v1.12.0-27B-i1-GGUF

NaNK
2,867
0

Cicikus-v3-1.4B-i1-GGUF

NaNK
llama-3.2
2,864
0

CodeV-R1-Qwen-7B-i1-GGUF

NaNK
2,855
3

Qwen-SEA-LION-v4-4B-VL-Magic_decensored-i1-GGUF

NaNK
2,855
1

Ministral-3-8B-Instruct-2512-tainted-heresy-i1-GGUF

NaNK
license:apache-2.0
2,855
1

reactor-ai-20b-i1-GGUF

NaNK
license:apache-2.0
2,852
0

Fyodor-Q3-8B-Instruct-i1-GGUF

NaNK
license:apache-2.0
2,826
1

sundae-v716-update-direct-4b-i1-GGUF

NaNK
license:cc-by-nc-4.0
2,820
0

ClinAligh-4B-i1-GGUF

NaNK
license:apache-2.0
2,819
1

Hunyuan-MT-Chimera-7B-i1-GGUF

weighted/imatrix quants of https://huggingface.co/tencent/Hunyuan-MT-Chimera-7B For a convenient overview and download list, visit our model page for this model. static quants are available at https://huggingface.co/mradermacher/Hunyuan-MT-Chimera-7B-GGUF Usage If you are unsure how to use GGUF files, refer to one of TheBloke's READMEs for more details, including on how to concatenate multi-part files. (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | GGUF | imatrix | 0.1 | imatrix file (for creating your own qwuants) | | GGUF | i1-IQ1S | 1.9 | for the desperate | | GGUF | i1-IQ1M | 2.1 | mostly desperate | | GGUF | i1-IQ2XXS | 2.3 | | | GGUF | i1-IQ2XS | 2.5 | | | GGUF | i1-IQ2S | 2.6 | | | GGUF | i1-IQ2M | 2.8 | | | GGUF | i1-Q2KS | 2.9 | very low quality | | GGUF | i1-Q2K | 3.1 | IQ3XXS probably better | | GGUF | i1-IQ3XXS | 3.1 | lower quality | | GGUF | i1-IQ3XS | 3.4 | | | GGUF | i1-Q3KS | 3.5 | IQ3XS probably better | | GGUF | i1-IQ3S | 3.6 | beats Q3K | | GGUF | i1-IQ3M | 3.7 | | | GGUF | i1-Q3KM | 3.9 | IQ3S probably better | | GGUF | i1-Q3KL | 4.2 | IQ3M probably better | | GGUF | i1-IQ4XS | 4.3 | | | GGUF | i1-Q40 | 4.5 | fast, low quality | | GGUF | i1-IQ4NL | 4.5 | prefer IQ4XS | | GGUF | i1-Q4KS | 4.5 | optimal size/speed/quality | | GGUF | i1-Q4KM | 4.7 | fast, recommended | | GGUF | i1-Q41 | 4.9 | | | GGUF | i1-Q5KS | 5.3 | | | GGUF | i1-Q5KM | 5.5 | | | GGUF | i1-Q6K | 6.3 | practically like static Q6K | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 See https://huggingface.co/mradermacher/modelrequests for some answers to questions you might have and/or if you want some other model quantized. I thank my company, nethype GmbH, for letting me use its servers and providing upgrades to my workstation to enable this work in my free time. Additional thanks to @nicoboss for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to.

NaNK
2,808
1

llama3-daybreak-lumimaid0.1-8b-hf-GGUF

NaNK
2,794
1

SADeepCoder-14B-Preview-unsloth-v1.0-i1-GGUF

NaNK
license:mit
2,794
0

Qwen3-4B-Thinking-2507-Claude-4.5-Opus-High-Reasoning-Distill-i1-GGUF

NaNK
license:apache-2.0
2,793
0

DR-Tulu-SFT-8B-i1-GGUF

NaNK
llama-factory
2,791
0

DiStil-Qwen3-1.7B-uncensored-i1-GGUF

NaNK
license:apache-2.0
2,790
0

Broken-Tutu-24B-Transgression-v2.0-GGUF

NaNK
license:apache-2.0
2,784
2

Qwen3-VL-8B-Medical-Extraction-i1-GGUF

NaNK
2,779
0

Olmo-3-7B-RL-Zero-IF-i1-GGUF

NaNK
license:apache-2.0
2,779
0

AutoGLM-Phone-9B-Multilingual-i1-GGUF

NaNK
license:mit
2,776
0

Qwen3.5-27B-Claude-4.6-Opus-Reasoning-Distilled-heretic-GGUF

NaNK
license:apache-2.0
2,769
2

Huihui-Qwen3-VL-32B-Instruct-abliterated-GGUF

static quants of https://huggingface.co/huihui-ai/Huihui-Qwen3-VL-32B-Instruct-abliterated For a convenient overview and download list, visit our model page for this model. weighted/imatrix quants are available at https://huggingface.co/mradermacher/Huihui-Qwen3-VL-32B-Instruct-abliterated-i1-GGUF Usage If you are unsure how to use GGUF files, refer to one of TheBloke's READMEs for more details, including on how to concatenate multi-part files. (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | GGUF | mmproj-Q80 | 0.9 | multi-modal supplement | | GGUF | mmproj-f16 | 1.3 | multi-modal supplement | | GGUF | Q2K | 12.4 | | | GGUF | Q3KS | 14.5 | | | GGUF | Q3KM | 16.1 | lower quality | | GGUF | Q3KL | 17.4 | | | GGUF | IQ4XS | 18.0 | | | GGUF | Q4KS | 18.9 | fast, recommended | | GGUF | Q4KM | 19.9 | fast, recommended | | GGUF | Q5KS | 22.7 | | | GGUF | Q5KM | 23.3 | | | GGUF | Q6K | 27.0 | very good quality | | GGUF | Q80 | 34.9 | fast, best quality | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 See https://huggingface.co/mradermacher/modelrequests for some answers to questions you might have and/or if you want some other model quantized. I thank my company, nethype GmbH, for letting me use its servers and providing upgrades to my workstation to enable this work in my free time.

NaNK
license:apache-2.0
2,765
1

SynLogic-7B-i1-GGUF

NaNK
license:mit
2,760
1

Qwen3-15B-A2B-Base-i1-GGUF

NaNK
license:apache-2.0
2,759
0

Sunflower-32B-ultravox-merged-ft-salt-instruct-i1-GGUF

NaNK
2,758
0

Qwen3.5-27B-heretic-v2-GGUF

NaNK
license:apache-2.0
2,757
9

Qwen3.5-27B-Claude-4.6-Opus-Reasoning-Distilled-heretic-v2-i1-GGUF

NaNK
license:apache-2.0
2,755
4

SynLogic-Mix-3-32B-i1-GGUF

NaNK
license:mit
2,753
0

Clado-BrowserOS-Action-i1-GGUF

license:apache-2.0
2,752
0

Huihui-Qwen3-VL-30B-A3B-Instruct-abliterated-i1-GGUF

weighted/imatrix quants of https://huggingface.co/huihui-ai/Huihui-Qwen3-VL-30B-A3B-Instruct-abliterated For a convenient overview and download list, visit our model page for this model. static quants are available at https://huggingface.co/mradermacher/Huihui-Qwen3-VL-30B-A3B-Instruct-abliterated-GGUF This is a vision model - mmproj files (if any) will be in the static repository. Usage If you are unsure how to use GGUF files, refer to one of TheBloke's READMEs for more details, including on how to concatenate multi-part files. (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | GGUF | imatrix | 0.2 | imatrix file (for creating your own qwuants) | | GGUF | i1-IQ1S | 6.5 | for the desperate | | GGUF | i1-IQ1M | 7.2 | mostly desperate | | GGUF | i1-IQ2XXS | 8.3 | | | GGUF | i1-IQ2XS | 9.2 | | | GGUF | i1-IQ2S | 9.4 | | | GGUF | i1-IQ2M | 10.3 | | | GGUF | i1-Q2KS | 10.6 | very low quality | | GGUF | i1-Q2K | 11.4 | IQ3XXS probably better | | GGUF | i1-IQ3XXS | 11.9 | lower quality | | GGUF | i1-IQ3XS | 12.7 | | | GGUF | i1-Q3KS | 13.4 | IQ3XS probably better | | GGUF | i1-IQ3S | 13.4 | beats Q3K | | GGUF | i1-IQ3M | 13.6 | | | GGUF | i1-Q3KM | 14.8 | IQ3S probably better | | GGUF | i1-Q3KL | 16.0 | IQ3M probably better | | GGUF | i1-IQ4XS | 16.5 | | | GGUF | i1-Q40 | 17.5 | fast, low quality | | GGUF | i1-Q4KS | 17.6 | optimal size/speed/quality | | GGUF | i1-Q4KM | 18.7 | fast, recommended | | GGUF | i1-Q41 | 19.3 | | | GGUF | i1-Q5KS | 21.2 | | | GGUF | i1-Q5KM | 21.8 | | | GGUF | i1-Q6K | 25.2 | practically like static Q6K | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 See https://huggingface.co/mradermacher/modelrequests for some answers to questions you might have and/or if you want some other model quantized. I thank my company, nethype GmbH, for letting me use its servers and providing upgrades to my workstation to enable this work in my free time. Additional thanks to @nicoboss for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to.

NaNK
license:apache-2.0
2,748
3

actio-ui-7b-sft-i1-GGUF

NaNK
2,747
1

Lang2Act-7B-i1-GGUF

NaNK
license:apache-2.0
2,747
0

LFM2-24B-A2B-abliterated-i1-GGUF

NaNK
license:apache-2.0
2,733
0

meteor-v4-2048-i1-GGUF

license:apache-2.0
2,731
1

gpt-oss-20b-gemini-2.5-pro-distill-GGUF

static quants of https://huggingface.co/armand0e/gpt-oss-20b-gemini-2.5-pro-distill For a convenient overview and download list, visit our model page for this model. weighted/imatrix quants are available at https://huggingface.co/mradermacher/gpt-oss-20b-gemini-2.5-pro-distill-i1-GGUF Usage If you are unsure how to use GGUF files, refer to one of TheBloke's READMEs for more details, including on how to concatenate multi-part files. (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | GGUF | Q3KS | 12.2 | | | GGUF | Q2K | 12.2 | | | GGUF | IQ4XS | 12.3 | | | GGUF | Q3KM | 13.0 | lower quality | | GGUF | Q3KL | 13.4 | | | GGUF | Q4KS | 14.8 | fast, recommended | | GGUF | Q4KM | 15.9 | fast, recommended | | GGUF | Q5KS | 16.0 | | | GGUF | Q5KM | 17.0 | | | GGUF | Q6K | 22.3 | very good quality | | GGUF | Q80 | 22.4 | fast, best quality | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 See https://huggingface.co/mradermacher/modelrequests for some answers to questions you might have and/or if you want some other model quantized. I thank my company, nethype GmbH, for letting me use its servers and providing upgrades to my workstation to enable this work in my free time.

NaNK
license:apache-2.0
2,724
0

Pelican1.0-VL-235B-A22B-FC-i1-GGUF

NaNK
license:apache-2.0
2,723
0

Disctil-Qwen3-1.7B-i1-GGUF

NaNK
2,721
0

MathSmith-hc-Qwen3-8B-i1-GGUF

NaNK
llama-factory
2,720
0

STAR1-R1-Distill-8B-i1-GGUF

NaNK
license:apache-2.0
2,717
0

AutoL2S-Plus-7b-i1-GGUF

NaNK
license:apache-2.0
2,702
0

Gemma-3-27B-Heretic-i1-GGUF

NaNK
2,695
0

Magidonia-24B-v4.3-creative-ORPO-v3-i1-GGUF

NaNK
2,691
2

Qwen-3.5-27B-Derestricted-GGUF

NaNK
unlimited
2,684
6

DeepSeek-R1-Distill-Qwen-1.5B-uncensored-GGUF

NaNK
license:mit
2,681
16

gemma-4-31B-it-Grand-Horror-X-INTENSE-HERETIC-UNCENSORED-Thinking-i1-GGUF

NaNK
license:apache-2.0
2,679
0

mistralai-Mistral-Nemo-Instruct-2407-12B-MPOA-v1-i1-GGUF

NaNK
license:apache-2.0
2,678
1

Llama-4-Scout-17B-16E-Instruct-abliterated-i1-GGUF

NaNK
llama
2,673
2

Huihui-gpt-oss-20b-BF16-abliterated-v2-GGUF

static quants of https://huggingface.co/huihui-ai/Huihui-gpt-oss-20b-BF16-abliterated-v2 For a convenient overview and download list, visit our model page for this model. weighted/imatrix quants are available at https://huggingface.co/mradermacher/Huihui-gpt-oss-20b-BF16-abliterated-v2-i1-GGUF Usage If you are unsure how to use GGUF files, refer to one of TheBloke's READMEs for more details, including on how to concatenate multi-part files. (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | GGUF | Q3KS | 12.2 | | | GGUF | Q2K | 12.2 | | | GGUF | IQ4XS | 12.3 | | | GGUF | Q3KM | 13.0 | lower quality | | GGUF | Q3KL | 13.4 | | | GGUF | Q4KS | 14.8 | fast, recommended | | GGUF | Q4KM | 15.9 | fast, recommended | | GGUF | Q5KS | 16.0 | | | GGUF | Q5KM | 17.0 | | | GGUF | Q6K | 22.3 | very good quality | | GGUF | Q80 | 22.4 | fast, best quality | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 See https://huggingface.co/mradermacher/modelrequests for some answers to questions you might have and/or if you want some other model quantized. I thank my company, nethype GmbH, for letting me use its servers and providing upgrades to my workstation to enable this work in my free time.

NaNK
license:apache-2.0
2,668
6

Seed-OSS-36B-Instruct-heretic-i1-GGUF

NaNK
license:apache-2.0
2,667
2

Ministral-3-3B-Instruct-2512-BF16-i1-GGUF

NaNK
license:apache-2.0
2,662
0

Olmo-3-1025-7B-i1-GGUF

NaNK
license:apache-2.0
2,659
0

GRiP-i1-GGUF

license:apache-2.0
2,653
1

RSI-AI-V1.1-GGUF

NaNK
license:apache-2.0
2,653
0

StrikeGPT-VL-8B-i1-GGUF

NaNK
license:apache-2.0
2,651
0

Kimi-VL-A3B-Thinking-2506-GGUF

NaNK
license:mit
2,650
2

Olmo-3-7B-Think-SFT-i1-GGUF

NaNK
license:apache-2.0
2,642
1

Llama3-8B-senator-i1-GGUF

NaNK
base_model:Weiyifan/Llama3-8B-senator
2,641
0

Nexura-Gemma2B-i1-GGUF

NaNK
2,639
0

L3.3-The-Omega-Directive-70B-Unslop-v2.0-GGUF

NaNK
license:llama3.3
2,637
1

Mira-v1.17-Karcher-27B-i1-GGUF

NaNK
2,634
0

CAI-20B-v2-i1-GGUF

NaNK
license:mit
2,629
0

Gemma3-Emotional-1B-i1-GGUF

NaNK
2,622
0

Orion-Qwen3.5-2B-SFT-v2603-v1-i1-GGUF

NaNK
license:apache-2.0
2,617
0

SwarmMed-14B-v1.2-merged-i1-GGUF

NaNK
license:apache-2.0
2,617
0

ATLAS-Teach-8B-Instruct-i1-GGUF

weighted/imatrix quants of https://huggingface.co/Arc-Intelligence/ATLAS-8B-Instruct For a convenient overview and download list, visit our model page for this model. static quants are available at https://huggingface.co/mradermacher/ATLAS-Teach-8B-Instruct-GGUF Usage If you are unsure how to use GGUF files, refer to one of TheBloke's READMEs for more details, including on how to concatenate multi-part files. (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | GGUF | imatrix | 0.1 | imatrix file (for creating your own qwuants) | | GGUF | i1-IQ1S | 2.2 | for the desperate | | GGUF | i1-IQ1M | 2.4 | mostly desperate | | GGUF | i1-IQ2XXS | 2.6 | | | GGUF | i1-IQ2XS | 2.8 | | | GGUF | i1-IQ2S | 3.0 | | | GGUF | i1-IQ2M | 3.2 | | | GGUF | i1-Q2KS | 3.2 | very low quality | | GGUF | i1-Q2K | 3.4 | IQ3XXS probably better | | GGUF | i1-IQ3XXS | 3.5 | lower quality | | GGUF | i1-IQ3XS | 3.7 | | | GGUF | i1-Q3KS | 3.9 | IQ3XS probably better | | GGUF | i1-IQ3S | 3.9 | beats Q3K | | GGUF | i1-IQ3M | 4.0 | | | GGUF | i1-Q3KM | 4.2 | IQ3S probably better | | GGUF | i1-Q3KL | 4.5 | IQ3M probably better | | GGUF | i1-IQ4XS | 4.7 | | | GGUF | i1-Q40 | 4.9 | fast, low quality | | GGUF | i1-IQ4NL | 4.9 | prefer IQ4XS | | GGUF | i1-Q4KS | 4.9 | optimal size/speed/quality | | GGUF | i1-Q4KM | 5.1 | fast, recommended | | GGUF | i1-Q41 | 5.3 | | | GGUF | i1-Q5KS | 5.8 | | | GGUF | i1-Q5KM | 6.0 | | | GGUF | i1-Q6K | 6.8 | practically like static Q6K | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 See https://huggingface.co/mradermacher/modelrequests for some answers to questions you might have and/or if you want some other model quantized. I thank my company, nethype GmbH, for letting me use its servers and providing upgrades to my workstation to enable this work in my free time. Additional thanks to @nicoboss for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to.

NaNK
license:apache-2.0
2,612
0

Lit-Qwen3-32B-i1-GGUF

NaNK
license:apache-2.0
2,612
0

turkish-llm-7b-instruct-i1-GGUF

NaNK
license:apache-2.0
2,611
0

SynLogic-32B-i1-GGUF

NaNK
license:mit
2,604
0

llm-surgery-dark-arts-gpt-oss-60b-96a12-i1-GGUF

NaNK
license:apache-2.0
2,603
0

Suri-Qwen-3.5-9B-Uncensored-Soft-i1-GGUF

NaNK
2,598
1

PAPO-G-Qwen2.5-VL-3B-i1-GGUF

NaNK
license:mit
2,595
0

llmfan46-Qwen3.5-9B-ultra-heretic-i1-GGUF

NaNK
license:apache-2.0
2,594
0

tvall43-Qwen3.5-4B-heretic-v2-i1-GGUF

NaNK
license:apache-2.0
2,591
1

Mira-v1.12-Ties-27B-i1-GGUF

NaNK
2,582
0