tensorblock

500 models • 1 total models in database
Sort by:

Llama-3.2-8B-Instruct-GGUF

NaNK
base_model:voidful/Llama-3.2-8B-Instruct
19,807
2

llama3.2-1b-Uncensored-GGUF

NaNK
llama
2,161
6

Phi-4-mini-instruct-abliterated-GGUF

license:mit
1,287
7

Qwen2.5-7B-Instruct-Uncensored-GGUF

NaNK
license:gpl-3.0
1,098
6

deepseek-coder-7b-instruct-v1.5-GGUF

NaNK
1,043
1

PocketDoc_Dans-PersonalityEngine-V1.3.0-12b-GGUF

NaNK
license:apache-2.0
914
1

Qwen2.5-32B-Instruct-abliterated-GGUF

NaNK
license:apache-2.0
848
3

mistral-7b-uncensored-GGUF

NaNK
license:apache-2.0
832
4

DarkIdol-Llama-3.1-8B-Instruct-1.2-Uncensored-GGUF

NaNK
llama3
815
5

Deepseek-R1-Distill-NSFW-RPv1-GGUF

NaNK
775
0

UnfilteredAI_DAN-Qwen3-1.7B-GGUF

NaNK
license:apache-2.0
758
5

WebSquareAI-Instruct-llama-3-8B-v0.5.39-GGUF

NaNK
base_model:kimwooglae/WebSquareAI-Instruct-llama-3-8B-v0.5.39
740
0

SwallowMaid-8B-L3-SPPO-abliterated-GGUF

NaNK
llama
728
0

llama3.1_korean_v0.1_sft_by_aidx-GGUF

base_model:SEOKDONG/llama3.1_korean_v0.1_sft_by_aidx
699
0

Qwen_Qwen3-Coder-30B-A3B-Instruct-GGUF

NaNK
license:apache-2.0
693
1

BioLlama-Ko-8B-GGUF

NaNK
base_model:iRASC/BioLlama-Ko-8B
673
0

Phi-3.5-mini-instruct-GGUF

license:mit
655
1

sexyGPT-Uncensored-GGUF

license:apache-2.0
649
7

Qwen2.5-3B-Instruct-GGUF

NaNK
637
1

Midnight-Miqu-70B-v1.5-GGUF

NaNK
611
1

Qwen2.5-3B-Instruct-Uncensored-Test-GGUF

NaNK
llama-factory
610
2

DeepSeek-R1-Distill-Llama-3B-GGUF

NaNK
base_model:suayptalha/DeepSeek-R1-Distill-Llama-3B
603
2

NemoMix-Unleashed-12B-GGUF

NaNK
600
0

llama3-8B-slerp-persian-merge-GGUF

NaNK
base_model:melino2000/llama3-8B-slerp-persian-merge
590
1

Llama-3.2-1B-Instruct-abliterated-GGUF

NaNK
base_model:huihui-ai/Llama-3.2-1B-Instruct-abliterated
562
1

gemma-2-9b-instruct-GGUF

NaNK
559
0

Meta-Llama-3.1-70B-Instruct-GGUF

NaNK
base_model:unsloth/Meta-Llama-3.1-70B-Instruct
551
0

Llama3-Aloe-8B-Alpha-GGUF

NaNK
base_model:HPAI-BSC/Llama3-Aloe-8B-Alpha
543
1

Llama-3.2-3B-Overthinker-GGUF

NaNK
llama
540
0

Llama-3.1-8B-GGUF

NaNK
llama
534
0

cyber-risk-llama-3-8b-instruct-sft-GGUF

NaNK
base_model:Vanessasml/cyber-risk-llama-3-8b-instruct-sft
521
0

ANIMA-Nectar-v2-GGUF

NaNK
license:mit
510
0

TowerInstruct-Mistral-7B-v0.2-GGUF

NaNK
license:cc-by-nc-4.0
498
2

gemma-3-4b-it-GGUF

NaNK
498
0

Xenova_bloom-560m-GGUF

472
0

MultiverseEx26-7B-slerp-GGUF

NaNK
license:apache-2.0
472
0

gpt4all-falcon-GGUF

license:apache-2.0
470
1

DeepSeek-R1-Qwen2.5-1.5b-SFT-R1-JSON-Unstructured-To-Structured-GGUF

NaNK
license:apache-2.0
460
0

dolphin-2.9.3-llama-3-8b-GGUF

NaNK
base_model:dphn/dolphin-2.9.3-llama-3-8b
456
1

Llama-3-uncensored-Dare-1-GGUF

NaNK
nbeerbower/llama-3-spicy-abliterated-stella-8B
448
0

llama3.2-3b-uncensored-GGUF

NaNK
base_model:thirdeyeai/llama3.2-3b-uncensored
439
0

Mixtral-8x7B-Instruct-v0.1-GGUF

NaNK
license:apache-2.0
432
1

L500MT-GGUF

license:apache-2.0
422
0

L3-8B-Stheno-v3.2-GGUF

NaNK
license:cc-by-nc-4.0
420
1

ReflectionCoder-DS-33B-GGUF

NaNK
license:apache-2.0
417
0

Calcium-Opus-14B-Elite-1M-GGUF

NaNK
license:apache-2.0
416
0

Yi-6B-200K-GGUF

NaNK
license:apache-2.0
414
0

Qwen2.5-7B-Instruct-GGUF

NaNK
license:apache-2.0
413
1

llama3-eng-ko-8b-sl-GGUF

NaNK
llama-3-ko
413
0

Llama-3-Instruct-8B-DPO-GGUF

NaNK
base_model:princeton-nlp/Llama-3-Instruct-8B-DPO
412
0

Llama-3.1-8B-Ko-bigdefence-GGUF

NaNK
llama
411
0

tokyotech-llm_Llama-3.1-Swallow-8B-Instruct-v0.5-GGUF

[](https://tensorblock.co) [](https://twitter.com/tensorblockaoi) [](https://discord.gg/Ej5NmeHFf2) [](https://github.com/TensorBlock) [](https://t.me/TensorBlock) tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.5 - GGUF Join our Discord to learn more about what we're building ↗ This repo contains GGUF format model files for tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.5. The files were quantized using machines provided by TensorBlock, and they are compatible with llama.cpp as of commit b5753. A comprehensive collection of Model Context Protocol (MCP) servers. A lightweight, open, and extensible multi-LLM interaction studio. | Filename | Quant type | File Size | Description | | -------- | ---------- | --------- | ----------- | | Llama-3.1-Swallow-8B-Instruct-v0.5-Q2K.gguf | Q2K | 3.179 GB | smallest, significant quality loss - not recommended for most purposes | | Llama-3.1-Swallow-8B-Instruct-v0.5-Q3KS.gguf | Q3KS | 3.665 GB | very small, high quality loss | | Llama-3.1-Swallow-8B-Instruct-v0.5-Q3KM.gguf | Q3KM | 4.019 GB | very small, high quality loss | | Llama-3.1-Swallow-8B-Instruct-v0.5-Q3KL.gguf | Q3KL | 4.322 GB | small, substantial quality loss | | Llama-3.1-Swallow-8B-Instruct-v0.5-Q40.gguf | Q40 | 4.661 GB | legacy; small, very high quality loss - prefer using Q3KM | | Llama-3.1-Swallow-8B-Instruct-v0.5-Q4KS.gguf | Q4KS | 4.693 GB | small, greater quality loss | | Llama-3.1-Swallow-8B-Instruct-v0.5-Q4KM.gguf | Q4KM | 4.921 GB | medium, balanced quality - recommended | | Llama-3.1-Swallow-8B-Instruct-v0.5-Q50.gguf | Q50 | 5.599 GB | legacy; medium, balanced quality - prefer using Q4KM | | Llama-3.1-Swallow-8B-Instruct-v0.5-Q5KS.gguf | Q5KS | 5.599 GB | large, low quality loss - recommended | | Llama-3.1-Swallow-8B-Instruct-v0.5-Q5KM.gguf | Q5KM | 5.733 GB | large, very low quality loss - recommended | | Llama-3.1-Swallow-8B-Instruct-v0.5-Q6K.gguf | Q6K | 6.596 GB | very large, extremely low quality loss | | Llama-3.1-Swallow-8B-Instruct-v0.5-Q80.gguf | Q80 | 8.541 GB | very large, extremely low quality loss - not recommended | Then, downoad the individual model file the a local directory If you wanna download multiple model files with a pattern (e.g., `Q4Kgguf`), you can try:

NaNK
base_model:tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.5
408
0

Llama-3-8B-Lexi-Uncensored-GGUF

NaNK
llama3
401
1

llama-3-debug-GGUF

llama-3
400
0

L200MT-GGUF

license:apache-2.0
400
0

bloomz-3b-GGUF

NaNK
398
0

gemma-3-12b-it-GGUF

NaNK
395
1

Mistral-Nemo-Instruct-2407-GGUF

license:apache-2.0
390
0

WhiteRabbitNeo-2.5-Qwen-2.5-Coder-7B-GGUF

NaNK
license:apache-2.0
389
2

Llama-Primus-Merged-GGUF

base_model:trendmicro-ailab/Llama-Primus-Merged
388
1

AceGPT-v2-8B-Chat-GGUF

NaNK
license:apache-2.0
387
0

llama-3-typhoon-v1.5x-8b-instruct-GGUF

NaNK
base_model:typhoon-ai/llama-3-typhoon-v1.5x-8b-instruct
385
1

gemma-2-9b-GGUF

NaNK
383
0

Fireball-Alpaca-Llama3.1.08-8B-Philos-C-R1-KTO-beta-GGUF

NaNK
llama
376
0

MagpieLM-8B-SFT-v0.1-GGUF

NaNK
license:llama3.1
374
0

Qwen2-7B-GGUF

NaNK
license:apache-2.0
373
0

calme-2.6-qwen2-7b-GGUF

NaNK
license:apache-2.0
373
0

Llama-3-Trendyol-LLM-8b-chat-v2.0-GGUF

NaNK
base_model:Trendyol/Llama-3-Trendyol-LLM-8b-chat-v2.0
372
0

Lexora-Lite-3B-GGUF

NaNK
365
1

rakeshkiriyath_gpt2Medium_text_to_sql-GGUF

362
1

SlimOrca-13B-GGUF

NaNK
license:cc-by-nc-nd-4.0
361
0

SeaQwen2-0.5B-GGUF

NaNK
license:apache-2.0
358
0

Pinkstack_Base-Roblox-coder-Llama-3.2-3B-vLLM-GGUF

NaNK
llama
355
0

llama2-13b-dpo-v4-GGUF

NaNK
base_model:mncai/llama2-13b-dpo-v4
355
0

Qwen1.5-7B-GGUF

NaNK
354
0

DeepSeek-R1-Distill-Llama-8B-abliterated-GGUF

NaNK
base_model:huihui-ai/DeepSeek-R1-Distill-Llama-8B-abliterated
352
1

Phi-3-mini-128k-instruct-GGUF

license:mit
352
0

Qra-13B-chat-GGUF

NaNK
llama
352
0

DeepSeek-R1-Strategy-Qwen-2.5-1.5b-Unstructured-To-Structured-GGUF

NaNK
license:apache-2.0
349
1

Configurable-Hermes-3-Llama-3.1-8B-GGUF

NaNK
base_model:vicgalle/Configurable-Hermes-3-Llama-3.1-8B
349
0

Qra-1b-GGUF

NaNK
license:apache-2.0
348
1

granite-8b-code-base-4k-GGUF

NaNK
license:apache-2.0
346
0

llama-3-youko-8b-instruct-GGUF

NaNK
llama
344
0

NemoReRemix-12B-GGUF

NaNK
343
0

tiny-llama3-test-GGUF

base_model:ariG23498/tiny-llama3-test
343
0

Llama-3.1-8B-Lexi-Uncensored-V2-GGUF

NaNK
base_model:Orenguteng/Llama-3.1-8B-Lexi-Uncensored-V2
341
3

DavidAU_Qwen3-8B-64k-Context-2X-Josiefied-Uncensored-GGUF

NaNK
340
3

DeepSeek-V3-1B-Test-GGUF

NaNK
license:mit
333
3

Llama-3-OpenBioMed-8B-slerp-v0.3-GGUF

NaNK
skumar9/Llama-medx_v3
332
0

SOLAR-10.7B-slerp-GGUF

NaNK
license:apache-2.0
332
0

jpacifico_French-Alpaca-Llama3-8B-Instruct-v1.0-GGUF

NaNK
llama3
331
0

starcoder2-15b-instruct-v0.1-GGUF

NaNK
328
0

mistral-7b-grok-GGUF

NaNK
license:apache-2.0
325
4

Llama-3-8B-Instruct-abliterated-v2-GGUF

NaNK
base_model:QuixiAI/Llama-3-8B-Instruct-abliterated-v2
323
1

mims-harvard_TxAgent-T1-Llama-3.1-8B-GGUF

NaNK
base_model:mims-harvard/TxAgent-T1-Llama-3.1-8B
323
0

OpenCrystal-12B-L3.1-128K-GGUF

NaNK
322
0

llama3-8B-DarkIdol-2.3-Uncensored-32K-GGUF

NaNK
llama3
321
2

SecurityLLM-GGUF

license:apache-2.0
321
2

MBZUAI-Paris_Atlas-Chat-2B-GGUF

NaNK
320
0

yodayo-ai_nephra_v1.0-GGUF

NaNK
license:llama3
320
0

salamandra-7b-instruct-GGUF

NaNK
license:apache-2.0
315
1

Llama-medx_v3.1-GGUF

NaNK
base_model:skumar9/Llama-medx_v3.1
314
0

MiniMaxAI_SynLogic-7B-GGUF

[](https://tensorblock.co) [](https://twitter.com/tensorblockaoi) [](https://discord.gg/Ej5NmeHFf2) [](https://github.com/TensorBlock) [](https://t.me/TensorBlock) Join our Discord to learn more about what we're building ↗ This repo contains GGUF format model files for MiniMaxAI/SynLogic-7B. The files were quantized using machines provided by TensorBlock, and they are compatible with llama.cpp as of commit b5753. A comprehensive collection of Model Context Protocol (MCP) servers. A lightweight, open, and extensible multi-LLM interaction studio. | Filename | Quant type | File Size | Description | | -------- | ---------- | --------- | ----------- | | SynLogic-7B-Q2K.gguf | Q2K | 3.016 GB | smallest, significant quality loss - not recommended for most purposes | | SynLogic-7B-Q3KS.gguf | Q3KS | 3.492 GB | very small, high quality loss | | SynLogic-7B-Q3KM.gguf | Q3KM | 3.808 GB | very small, high quality loss | | SynLogic-7B-Q3KL.gguf | Q3KL | 4.088 GB | small, substantial quality loss | | SynLogic-7B-Q40.gguf | Q40 | 4.431 GB | legacy; small, very high quality loss - prefer using Q3KM | | SynLogic-7B-Q4KS.gguf | Q4KS | 4.458 GB | small, greater quality loss | | SynLogic-7B-Q4KM.gguf | Q4KM | 4.683 GB | medium, balanced quality - recommended | | SynLogic-7B-Q50.gguf | Q50 | 5.315 GB | legacy; medium, balanced quality - prefer using Q4KM | | SynLogic-7B-Q5KS.gguf | Q5KS | 5.315 GB | large, low quality loss - recommended | | SynLogic-7B-Q5KM.gguf | Q5KM | 5.445 GB | large, very low quality loss - recommended | | SynLogic-7B-Q6K.gguf | Q6K | 6.254 GB | very large, extremely low quality loss | | SynLogic-7B-Q80.gguf | Q80 | 8.099 GB | very large, extremely low quality loss - not recommended | Then, downoad the individual model file the a local directory If you wanna download multiple model files with a pattern (e.g., `Q4Kgguf`), you can try:

NaNK
license:mit
313
2

Ko-PlatYi-6B-kiwi-GGUF

NaNK
license:cc-by-nc-sa-4.0
309
0

defog_sqlcoder2-GGUF

NaNK
305
0

L100MT-GGUF

license:apache-2.0
303
0

scb10x_typhoon2.1-gemma3-12b-GGUF

[](https://tensorblock.co) [](https://twitter.com/tensorblockaoi) [](https://discord.gg/Ej5NmeHFf2) [](https://github.com/TensorBlock) [](https://t.me/TensorBlock) Join our Discord to learn more about what we're building ↗ This repo contains GGUF format model files for scb10x/typhoon2.1-gemma3-12b. The files were quantized using machines provided by TensorBlock, and they are compatible with llama.cpp as of commit b5753. A comprehensive collection of Model Context Protocol (MCP) servers. A lightweight, open, and extensible multi-LLM interaction studio. | Filename | Quant type | File Size | Description | | -------- | ---------- | --------- | ----------- | | typhoon2.1-gemma3-12b-Q2K.gguf | Q2K | 4.768 GB | smallest, significant quality loss - not recommended for most purposes | | typhoon2.1-gemma3-12b-Q3KS.gguf | Q3KS | 5.458 GB | very small, high quality loss | | typhoon2.1-gemma3-12b-Q3KM.gguf | Q3KM | 6.009 GB | very small, high quality loss | | typhoon2.1-gemma3-12b-Q3KL.gguf | Q3KL | 6.480 GB | small, substantial quality loss | | typhoon2.1-gemma3-12b-Q40.gguf | Q40 | 6.887 GB | legacy; small, very high quality loss - prefer using Q3KM | | typhoon2.1-gemma3-12b-Q4KS.gguf | Q4KS | 6.935 GB | small, greater quality loss | | typhoon2.1-gemma3-12b-Q4KM.gguf | Q4KM | 7.301 GB | medium, balanced quality - recommended | | typhoon2.1-gemma3-12b-Q50.gguf | Q50 | 8.232 GB | legacy; medium, balanced quality - prefer using Q4KM | | typhoon2.1-gemma3-12b-Q5KS.gguf | Q5KS | 8.232 GB | large, low quality loss - recommended | | typhoon2.1-gemma3-12b-Q5KM.gguf | Q5KM | 8.445 GB | large, very low quality loss - recommended | | typhoon2.1-gemma3-12b-Q6K.gguf | Q6K | 9.661 GB | very large, extremely low quality loss | | typhoon2.1-gemma3-12b-Q80.gguf | Q80 | 12.510 GB | very large, extremely low quality loss - not recommended | Then, downoad the individual model file the a local directory If you wanna download multiple model files with a pattern (e.g., `Q4Kgguf`), you can try:

NaNK
302
0

AMD-Llama-135m-code-GGUF

NaNK
base_model:amd/AMD-Llama-135m-code
295
0

Qwen2-VL-7B-Instruct-GGUF

NaNK
license:apache-2.0
295
0

dictalm2.0-instruct-GGUF

license:apache-2.0
294
0

tabula-8b-GGUF

NaNK
license:llama3
292
0

Aira-2-1B1-GGUF

NaNK
license:apache-2.0
292
0

DeepSeek-R1-Distill-Qwen-7B-abliterated-v2-GGUF

NaNK
290
3

Llama-3.2-1B-Instruct-GGUF

NaNK
llama
290
2

Infinirc-Llama3-8B-2G-Release-v1.0-GGUF

NaNK
base_model:Infinirc/Infinirc-Llama3-8B-2G-Release-v1.0
289
0

open-llama-3.2-1B-Instruct-GGUF

NaNK
base_model:diabolic6045/open-llama-3.2-1B-Instruct
286
0

L3.1-8B-sunfall-stheno-v0.6.1-GGUF

NaNK
license:llama3
285
2

distilgpt2-GGUF

NaNK
license:apache-2.0
285
0

SauerkrautLM-1.5b-GGUF

NaNK
license:apache-2.0
284
0

context_tuned_patient_matching_Llama-3.2-1B-Instruct-GGUF

NaNK
base_model:NAM00/context_tuned_patient_matching_Llama-3.2-1B-Instruct
283
0

SOLAR-10.7B-v1.1-GGUF

NaNK
license:apache-2.0
283
0

Python-Code-13B-GGUF

NaNK
license:cc-by-nc-nd-4.0
282
0

llama3-koen-sft-dpo-v1-GGUF

NaNK
llama-factory
281
0

Meta-Llama-3.1-8B-Instruct-abliterated-GGUF

NaNK
base_model:mlabonne/Meta-Llama-3.1-8B-Instruct-abliterated
281
0

Colibri_8b_v0.1-GGUF

NaNK
llama3
280
0

AMD-Llama-135m-GGUF

NaNK
base_model:amd/AMD-Llama-135m
280
0

LLaMA-Mesh-GGUF

base_model:Zhengyi/LLaMA-Mesh
278
1

Vikhr-Gemma-2B-instruct-GGUF

NaNK
license:apache-2.0
277
0

granite-20b-code-base-8k-GGUF

NaNK
license:apache-2.0
277
0

QwQ-32B-GGUF

NaNK
license:apache-2.0
276
4

DeepSeek-R1-Distill-Qwen-32B-abliterated-GGUF

NaNK
275
3

MultiLora-drop-sharegpt-GGUF

license:llama2
273
0

DCFT-Stratos-Unverified-114k-32B-GGUF

NaNK
llama-factory
272
0

deepseek-r1-14b-cot-math-reasoning-full-GGUF

NaNK
license:mit
268
2

Llama-3-13B-Instruct-GGUF

NaNK
base_model:elinas/Llama-3-13B-Instruct
268
1

gemma-2b-it-GGUF

NaNK
license:apache-2.0
268
0

gpt-neox-20b-GGUF

NaNK
license:apache-2.0
267
1

semi_final_Bllossom-GGUF

llama-factory
267
0

Llama-3-70B-Synthia-v3.5-GGUF

NaNK
base_model:migtissera/Llama-3-70B-Synthia-v3.5
267
0

zeta-GGUF

license:apache-2.0
265
1

BgGPT-Gemma-2-2.6B-IT-v1.0-GGUF

NaNK
265
0

Unbabel_TowerInstruct-13B-v0.1-GGUF

NaNK
license:cc-by-nc-4.0
265
0

llama-3-sqlcoder-8b-GGUF

NaNK
base_model:defog/llama-3-sqlcoder-8b
264
1

llama_16bit_2-GGUF

NaNK
llama
264
0

OpenR1-Qwen-7B-French-GGUF

NaNK
license:apache-2.0
263
1

starcoder2-7b-GGUF

NaNK
262
2

deepseek-coder-1.3b-instruct-GGUF

NaNK
262
2

granite-34b-code-base-8k-GGUF

NaNK
license:apache-2.0
262
0

kyx0r_Neona-12B-GGUF

[](https://tensorblock.co) [](https://twitter.com/tensorblockaoi) [](https://discord.gg/Ej5NmeHFf2) [](https://github.com/TensorBlock) [](https://t.me/TensorBlock) Join our Discord to learn more about what we're building ↗ This repo contains GGUF format model files for kyx0r/Neona-12B. The files were quantized using machines provided by TensorBlock, and they are compatible with llama.cpp as of commit b5753. A comprehensive collection of Model Context Protocol (MCP) servers. A lightweight, open, and extensible multi-LLM interaction studio. | Filename | Quant type | File Size | Description | | -------- | ---------- | --------- | ----------- | | Neona-12B-Q2K.gguf | Q2K | 4.791 GB | smallest, significant quality loss - not recommended for most purposes | | Neona-12B-Q3KS.gguf | Q3KS | 5.534 GB | very small, high quality loss | | Neona-12B-Q3KM.gguf | Q3KM | 6.083 GB | very small, high quality loss | | Neona-12B-Q3KL.gguf | Q3KL | 6.562 GB | small, substantial quality loss | | Neona-12B-Q40.gguf | Q40 | 7.072 GB | legacy; small, very high quality loss - prefer using Q3KM | | Neona-12B-Q4KS.gguf | Q4KS | 7.120 GB | small, greater quality loss | | Neona-12B-Q4KM.gguf | Q4KM | 7.477 GB | medium, balanced quality - recommended | | Neona-12B-Q50.gguf | Q50 | 8.519 GB | legacy; medium, balanced quality - prefer using Q4KM | | Neona-12B-Q5KS.gguf | Q5KS | 8.519 GB | large, low quality loss - recommended | | Neona-12B-Q5KM.gguf | Q5KM | 8.728 GB | large, very low quality loss - recommended | | Neona-12B-Q6K.gguf | Q6K | 10.056 GB | very large, extremely low quality loss | | Neona-12B-Q80.gguf | Q80 | 13.022 GB | very large, extremely low quality loss - not recommended | Then, downoad the individual model file the a local directory If you wanna download multiple model files with a pattern (e.g., `Q4Kgguf`), you can try:

NaNK
261
1

MN-DARKEST-UNIVERSE-29B-GGUF

NaNK
261
0

Monarch-7B-GGUF

NaNK
license:cc-by-nc-4.0
261
0

Llama-3-Alpha-Ko-8B-Instruct-GGUF

NaNK
base_model:allganize/Llama-3-Alpha-Ko-8B-Instruct
260
0

Oolel-v0.1-GGUF

NaNK
license:apache-2.0
258
0

Half-NSFW_Noromaid-7b-GGUF

NaNK
license:cc-by-4.0
258
0

Qwen-uncensored-v2-GGUF

NaNK
license:apache-2.0
257
2

Sailor-7B-GGUF

NaNK
license:apache-2.0
257
0

Fatgirl_v2_8B-GGUF

NaNK
257
0

FastApply-1.5B-v1.0-GGUF

NaNK
license:apache-2.0
257
0

deepseek-coder-6.7b-instruct-GGUF

NaNK
256
1

SuperNeuralDreadDevil-8b-GGUF

NaNK
256
1

internlm2_5-1_8b-chat-GGUF

NaNK
256
0

Qwen2.5-Coder-32B-Instruct-abliterated-GGUF

NaNK
license:apache-2.0
255
3

SmolLM2-360M-GGUF

license:apache-2.0
254
0

Viper-Coder-Hybrid-v1.3-GGUF

NaNK
license:apache-2.0
249
1

Llama-OpenReviewer-8B-GGUF

NaNK
base_model:maxidl/Llama-OpenReviewer-8B
249
0

Mistral-Small3-24B-InstructContinuedFine-GGUF

NaNK
license:apache-2.0
249
0

OLMoE-1B-7B-0924-GGUF

NaNK
license:apache-2.0
248
0

JungZoona_T3Q-qwen2.5-14b-v1.0-e3-GGUF

NaNK
license:apache-2.0
247
2

33x-coder-GGUF

license:apache-2.0
247
0

archangel_sft_llama7b-GGUF

NaNK
base_model:ContextualAI/archangel_sft_llama7b
247
0

AceInstruct-1.5B-GGUF

NaNK
license:cc-by-nc-4.0
247
0

DeepSeek-R1-Distill-Qwen-1.5B-GGUF

NaNK
license:apache-2.0
245
3

mxbai-rerank-large-v2-GGUF

NaNK
license:apache-2.0
245
2

stablelm-2-12b-GGUF

NaNK
245
0

MunicipalPredictionModel-Llama3-GGUF

NaNK
llama
245
0

huihui-ai_Huihui-Qwen3-4B-abliterated-v2-GGUF

NaNK
license:apache-2.0
245
0

Intelligent-Internet_II-Medical-8B-GGUF

[](https://tensorblock.co) [](https://twitter.com/tensorblockaoi) [](https://discord.gg/Ej5NmeHFf2) [](https://github.com/TensorBlock) [](https://t.me/TensorBlock) Join our Discord to learn more about what we're building ↗ This repo contains GGUF format model files for Intelligent-Internet/II-Medical-8B. The files were quantized using machines provided by TensorBlock, and they are compatible with llama.cpp as of commit b5753. A comprehensive collection of Model Context Protocol (MCP) servers. A lightweight, open, and extensible multi-LLM interaction studio. | Filename | Quant type | File Size | Description | | -------- | ---------- | --------- | ----------- | | II-Medical-8B-Q2K.gguf | Q2K | 3.282 GB | smallest, significant quality loss - not recommended for most purposes | | II-Medical-8B-Q3KS.gguf | Q3KS | 3.770 GB | very small, high quality loss | | II-Medical-8B-Q3KM.gguf | Q3KM | 4.124 GB | very small, high quality loss | | II-Medical-8B-Q3KL.gguf | Q3KL | 4.431 GB | small, substantial quality loss | | II-Medical-8B-Q40.gguf | Q40 | 4.775 GB | legacy; small, very high quality loss - prefer using Q3KM | | II-Medical-8B-Q4KS.gguf | Q4KS | 4.802 GB | small, greater quality loss | | II-Medical-8B-Q4KM.gguf | Q4KM | 5.028 GB | medium, balanced quality - recommended | | II-Medical-8B-Q50.gguf | Q50 | 5.721 GB | legacy; medium, balanced quality - prefer using Q4KM | | II-Medical-8B-Q5KS.gguf | Q5KS | 5.721 GB | large, low quality loss - recommended | | II-Medical-8B-Q5KM.gguf | Q5KM | 5.851 GB | large, very low quality loss - recommended | | II-Medical-8B-Q6K.gguf | Q6K | 6.726 GB | very large, extremely low quality loss | | II-Medical-8B-Q80.gguf | Q80 | 8.710 GB | very large, extremely low quality loss - not recommended | Then, downoad the individual model file the a local directory If you wanna download multiple model files with a pattern (e.g., `Q4Kgguf`), you can try:

NaNK
license:apache-2.0
244
1

gpt2-demo-GGUF

243
0

tinyllama-15M-GGUF

base_model:nickypro/tinyllama-15M
242
0

zephyr-python-ru-merged-GGUF

license:mit
242
0

Qwen1.5-0.5B-vortex-GGUF

NaNK
241
0

Dolphin3.0-Mistral-24B-GGUF

NaNK
241
0

mamba-2.8b-hf-GGUF

NaNK
239
0

ReadyArt_Broken-Tutu-24B-Unslop-v2.0-GGUF

NaNK
license:apache-2.0
238
1

luckychao_Vicuna-Backdoored-7B-GGUF

NaNK
236
0

Sirius-10B-GGUF

NaNK
license:apache-2.0
236
0

Qwen2.5-Coder-1.5B-GGUF

NaNK
license:apache-2.0
235
1

Aira-2-774M-GGUF

license:apache-2.0
234
0

beril-GGUF

234
0

Octopus-v2-GGUF

NaNK
license:cc-by-nc-4.0
233
0

SmolLM2-1.7B-GGUF

NaNK
license:apache-2.0
232
0

llama-3.2-1B-spinquant-hf-GGUF

NaNK
base_model:Hjgugugjhuhjggg/llama-3.2-1B-spinquant-hf
231
0

suzume-llama-3-8B-multilingual-orpo-borda-half-GGUF

NaNK
base_model:lightblue/suzume-llama-3-8B-multilingual-orpo-borda-half
230
1

Qwen1.5-7B-Chat-GGUF

NaNK
230
0

SherlockAssistant_Mistral-7B-Instruct-Ukrainian-GGUF

NaNK
license:apache-2.0
229
3

Guanaco-3B-Uncensored-v2-GGUF

NaNK
license:apache-2.0
229
1

Mistral-7B-Instruct-v0.2-GGUF

NaNK
license:apache-2.0
229
0

gemma-2-2b-neogenesis-ita-GGUF

NaNK
228
1

Qwen_Qwen3-1.7B-MLX-bf16-GGUF

[](https://tensorblock.co) [](https://twitter.com/tensorblockaoi) [](https://discord.gg/Ej5NmeHFf2) [](https://github.com/TensorBlock) [](https://t.me/TensorBlock) Join our Discord to learn more about what we're building ↗ This repo contains GGUF format model files for Qwen/Qwen3-1.7B-MLX-bf16. The files were quantized using machines provided by TensorBlock, and they are compatible with llama.cpp as of commit b5753. A comprehensive collection of Model Context Protocol (MCP) servers. A lightweight, open, and extensible multi-LLM interaction studio. | Filename | Quant type | File Size | Description | | -------- | ---------- | --------- | ----------- | | Qwen3-1.7B-MLX-bf16-Q2K.gguf | Q2K | 0.778 GB | smallest, significant quality loss - not recommended for most purposes | | Qwen3-1.7B-MLX-bf16-Q3KS.gguf | Q3KS | 0.867 GB | very small, high quality loss | | Qwen3-1.7B-MLX-bf16-Q3KM.gguf | Q3KM | 0.940 GB | very small, high quality loss | | Qwen3-1.7B-MLX-bf16-Q3KL.gguf | Q3KL | 1.004 GB | small, substantial quality loss | | Qwen3-1.7B-MLX-bf16-Q40.gguf | Q40 | 1.054 GB | legacy; small, very high quality loss - prefer using Q3KM | | Qwen3-1.7B-MLX-bf16-Q4KS.gguf | Q4KS | 1.060 GB | small, greater quality loss | | Qwen3-1.7B-MLX-bf16-Q4KM.gguf | Q4KM | 1.107 GB | medium, balanced quality - recommended | | Qwen3-1.7B-MLX-bf16-Q50.gguf | Q50 | 1.231 GB | legacy; medium, balanced quality - prefer using Q4KM | | Qwen3-1.7B-MLX-bf16-Q5KS.gguf | Q5KS | 1.231 GB | large, low quality loss - recommended | | Qwen3-1.7B-MLX-bf16-Q5KM.gguf | Q5KM | 1.258 GB | large, very low quality loss - recommended | | Qwen3-1.7B-MLX-bf16-Q6K.gguf | Q6K | 1.418 GB | very large, extremely low quality loss | | Qwen3-1.7B-MLX-bf16-Q80.gguf | Q80 | 1.834 GB | very large, extremely low quality loss - not recommended | Then, downoad the individual model file the a local directory If you wanna download multiple model files with a pattern (e.g., `Q4Kgguf`), you can try:

NaNK
license:apache-2.0
228
0

chat-gpt2-GGUF

NaNK
227
0

gpt2-GGUF

NaNK
license:mit
225
0

PiVoT-MoE-GGUF

license:cc-by-nc-4.0
222
0

QwQ-32B-bf16-GGUF

NaNK
license:apache-2.0
221
0

Meta-Llama-3-8B-Instruct-abliterated-v3-GGUF

NaNK
base_model:failspy/Meta-Llama-3-8B-Instruct-abliterated-v3
221
0

T-lite-instruct-0.1-GGUF

NaNK
219
0

semcoder_s_1030-GGUF

NaNK
219
0

granite-guardian-3.0-2b-GGUF

NaNK
license:apache-2.0
219
0

Delcos_Mistral-Pygmalion-7b-GGUF

NaNK
llama-2
218
0

Violet_Twilight-v0.2-GGUF

license:apache-2.0
218
0

qwen25-math-7b-instruct-GGUF

NaNK
license:apache-2.0
217
0

calme-2.4-qwen2-7b-GGUF

NaNK
license:apache-2.0
217
0

ross-dev_sexyGPT-Uncensored-GGUF

license:apache-2.0
215
2

Phi-4-mini-instruct-GGUF

license:mit
215
0

gemma-ko-1.1-2b-it-GGUF

NaNK
214
0

Llama-3.1-Nemotron-Nano-8B-v1-GGUF

NaNK
llama-3
213
6

Qwen2.5-Coder-14B-Instruct-abliterated-GGUF

NaNK
license:apache-2.0
213
0

granite-3.1-2b-instruct-GGUF

NaNK
license:apache-2.0
213
0

SWE-Fixer-Retriever-7B-GGUF

NaNK
license:mit
213
0

defog_sqlcoder-7b-GGUF

NaNK
license:cc-by-sa-4.0
212
0

kosolra_SFT_DPO_v0-GGUF

NaNK
license:mit
212
0

s1.1-7B-GGUF

NaNK
211
1

DeepSeek-R1-DRAFT-Qwen2.5-0.5B-GGUF

NaNK
license:apache-2.0
211
1

OpenR1-Qwen-7B-Turkish-GGUF

NaNK
license:apache-2.0
211
0

deepseek-coder-33b-instruct-GGUF

NaNK
211
0

Saul-7B-Instruct-v1-GGUF

NaNK
license:mit
210
0

SauerkrautLM-Gemma-7b-GGUF

NaNK
210
0

SlimMelodicMaid-GGUF

license:cc-by-4.0
210
0

MFANN-llama3.1-Abliterated-SLERP-GGUF

base_model:netcat420/MFANN-llama3.1-Abliterated-SLERP
208
0

CohereLabs_aya-23-8B-GGUF

NaNK
license:cc-by-nc-4.0
207
0

deepseek-math-7b-instruct-GGUF

NaNK
207
0

gemma-3-1b-it-abliterated-GGUF

NaNK
205
1

CHEMLLM-2b-1_5-GGUF

NaNK
license:apache-2.0
205
0

Llama-Song-Stream-3B-Instruct-GGUF

NaNK
Llama3.2
205
0

mistral_7b_0-3_oh-dcft-v3.1-claude-3-5-sonnet-20241022-GGUF

NaNK
llama-factory
204
1

ZeroAgency_Mistral-Small-3.1-24B-Instruct-2503-hf-GGUF

NaNK
license:apache-2.0
204
0

ghost-8b-beta-1608-GGUF

NaNK
llama
204
0

Indic-gemma-7b-finetuned-sft-Navarasa-2.0-GGUF

NaNK
204
0

cognitivecomputations_WizardLM-33B-V1.0-Uncensored-GGUF

NaNK
203
1

dolphincoder-starcoder2-15b-GGUF

NaNK
203
0

Triunvirato-7b-GGUF

NaNK
license:apache-2.0
203
0

Qwen2-VL-7B-GGUF

NaNK
license:apache-2.0
202
1

ArliAI_Qwen3-30B-A3B-ArliAI-RpR-v4-Fast-GGUF

NaNK
license:apache-2.0
201
0

DeepSeek-R1-Distill-Qwen-32B-GGUF

NaNK
license:apache-2.0
200
1

macbert4mdcspell_v1-GGUF

NaNK
license:apache-2.0
199
0

mistral-7b-dpo-v6-GGUF

NaNK
license:apache-2.0
199
0

SEOKDONG_llama3.1_korean_v1.1_sft_by_aidx-GGUF

llama3.1
199
0

s1k-GGUF

199
0

TeenyTinyLlama-160m-GGUF

base_model:nicholasKluge/TeenyTinyLlama-160m
198
0

Yi-34B-200K-DARE-megamerge-v8-GGUF

NaNK
197
0

gemma-2-27b-it-abliterated-GGUF

NaNK
196
3

llm4decompile-6.7b-v2-GGUF

NaNK
license:mit
196
0

karakuri-lm-8x7b-chat-v0.1-GGUF

NaNK
license:apache-2.0
196
0

Llama-3.1-Instruct_NSFW-pretrained_e1-plus_reddit-GGUF

llama
196
0

Mistral-Small-24B-Instruct-2501-abliterated-GGUF

NaNK
license:apache-2.0
195
2

TinyLlama-1.1B-32k-Instruct-GGUF

NaNK
llama
195
0

Tess-2.0-Llama-3-70B-GGUF

NaNK
base_model:migtissera/Tess-2.0-Llama-3-70B
195
0

mosaicml_mpt-7b-chat-GGUF

NaNK
license:cc-by-nc-sa-4.0
194
0

Llama-3.1-8B-Ultra-Instruct-GGUF

NaNK
base_model:Dampfinchen/Llama-3.1-8B-Ultra-Instruct
194
0

Lucie-7B-Instruct-v1.1-GGUF

NaNK
license:apache-2.0
194
0

Qwen2.5-Coder-7B-Instruct-GGUF

NaNK
license:apache-2.0
193
2

yamatazen_EtherealAurora-12B-v2-GGUF

NaNK
193
1

mohammedbriman_llama-2-7b-chat-turkish-instructions-GGUF

NaNK
base_model:mohammedbriman/llama-2-7b-chat-turkish-instructions
193
0

gpt2-650k-stable-diffusion-prompt-generator-GGUF

license:mit
192
2

Python-Code-33B-GGUF

NaNK
license:cc-by-nc-nd-4.0
192
0

Llama-3-instruction-constructionsafety-layertuning-GGUF

llama3
192
0

NSFW_DPO_Noromaid-7b-Mistral-7B-Instruct-v0.1-GGUF

NaNK
license:cc-by-nc-4.0
191
1

mlx-community_DeepSeek-R1-0528-Qwen3-8B-bf16-GGUF

NaNK
license:mit
190
0

mrfakename_mistral-small-3.1-24b-base-2503-hf-GGUF

NaNK
license:apache-2.0
190
0

nvidia_AceMath-RL-Nemotron-7B-GGUF

NaNK
190
0

llama-3-8b-gpt-4o-ru1.0-GGUF

NaNK
base_model:ruslandev/llama-3-8b-gpt-4o-ru1.0
188
1

chat_gpt2_dpo-GGUF

license:apache-2.0
188
0

s1.1-14B-GGUF

NaNK
188
0

prithivMLmods_Ophiuchi-Qwen3-14B-Instruct-GGUF

NaNK
license:apache-2.0
187
0

Roleplay-Llama-3-8B-GGUF

NaNK
base_model:vicgalle/Roleplay-Llama-3-8B
186
1

Wayfarer-Large-70B-Llama-3.3-GGUF

NaNK
base_model:LatitudeGames/Wayfarer-Large-70B-Llama-3.3
186
1

DeepSeek-R1-Distill-Qwen-7B-GGUF

NaNK
185
0

DeepSeek-R1-Distill-Llama-70B-abliterated-GGUF

NaNK
base_model:huihui-ai/DeepSeek-R1-Distill-Llama-70B-abliterated
185
0

tnayaj-GGUF

license:apache-2.0
185
0

h2ogpt-4096-llama2-7b-chat-GGUF

NaNK
llama
184
0

Qwen2-VL-2B-GGUF

NaNK
license:apache-2.0
183
0

TinyLlama-1.1B-Chat-v1.0-GGUF

NaNK
base_model:TinyLlama/TinyLlama-1.1B-Chat-v1.0
183
0

VityaVitalich_Llama3.1-8b-instruct-GGUF

NaNK
base_model:VityaVitalich/Llama3.1-8b-instruct
182
0

Q2AW1M-1100-GGUF

NaNK
license:apache-2.0
182
0

shieldgemma-2b-GGUF

NaNK
182
0

Qwen_Qwen3-0.6B-GGUF

NaNK
license:apache-2.0
181
3

gemma-ko-7b-GGUF

NaNK
181
0

shanghong_stage1-GGUF

[](https://tensorblock.co) [](https://twitter.com/tensorblockaoi) [](https://discord.gg/Ej5NmeHFf2) [](https://github.com/TensorBlock) [](https://t.me/TensorBlock) Join our Discord to learn more about what we're building ↗ This repo contains GGUF format model files for shanghong/stage1. The files were quantized using machines provided by TensorBlock, and they are compatible with llama.cpp as of commit b5753. A comprehensive collection of Model Context Protocol (MCP) servers. A lightweight, open, and extensible multi-LLM interaction studio. | Filename | Quant type | File Size | Description | | -------- | ---------- | --------- | ----------- | | stage1-Q2K.gguf | Q2K | 3.179 GB | smallest, significant quality loss - not recommended for most purposes | | stage1-Q3KS.gguf | Q3KS | 3.665 GB | very small, high quality loss | | stage1-Q3KM.gguf | Q3KM | 4.019 GB | very small, high quality loss | | stage1-Q3KL.gguf | Q3KL | 4.322 GB | small, substantial quality loss | | stage1-Q40.gguf | Q40 | 4.661 GB | legacy; small, very high quality loss - prefer using Q3KM | | stage1-Q4KS.gguf | Q4KS | 4.693 GB | small, greater quality loss | | stage1-Q4KM.gguf | Q4KM | 4.921 GB | medium, balanced quality - recommended | | stage1-Q50.gguf | Q50 | 5.599 GB | legacy; medium, balanced quality - prefer using Q4KM | | stage1-Q5KS.gguf | Q5KS | 5.599 GB | large, low quality loss - recommended | | stage1-Q5KM.gguf | Q5KM | 5.733 GB | large, very low quality loss - recommended | | stage1-Q6K.gguf | Q6K | 6.596 GB | very large, extremely low quality loss | | stage1-Q80.gguf | Q80 | 8.541 GB | very large, extremely low quality loss - not recommended | Then, downoad the individual model file the a local directory If you wanna download multiple model files with a pattern (e.g., `Q4Kgguf`), you can try:

NaNK
181
0

llama2-12.8b_lora-dpo_v1-GGUF

NaNK
base_model:etri-xainlp/llama2-12.8b_lora-dpo_v1
180
0

YuE-s1-7B-anneal-en-cot-GGUF

NaNK
license:cc-by-nc-4.0
179
28

sapie1-GGUF

NaNK
license:mit
179
0

Tifa-Deepsex-14b-CoT-GGUF

NaNK
license:apache-2.0
178
6

Qwen2.5-3B-Instruct-abliterated-GGUF

NaNK
license:apache-2.0
178
0

TinyMistral-6x248M-GGUF

license:apache-2.0
178
0

bloom-3b-conversational-GGUF

NaNK
license:mit
177
2

HelpingAI-3-GGUF

NaNK
177
1

0x-YuAN_codeparrot-ds-GGUF

175
0

Llama-3.2-3B-Instruct-uncensored-GGUF

NaNK
base_model:chuanli11/Llama-3.2-3B-Instruct-uncensored
175
0

LumiOpen_Llama-Poro-2-8B-Instruct-GGUF

NaNK
base_model:LumiOpen/Llama-Poro-2-8B-Instruct
175
0

YuE-s1-7B-anneal-zh-icl-GGUF

NaNK
license:cc-by-nc-4.0
174
1

cotran2_gemma3-1b-GGUF

NaNK
llama-factory
174
0

hivaze_ParaLex-Llama-3-8B-SFT-GGUF

NaNK
base_model:hivaze/ParaLex-Llama-3-8B-SFT
174
0

SakanaAI_Llama-3-8B-Instruct-Coding-Expert-GGUF

NaNK
base_model:SakanaAI/Llama-3-8B-Instruct-Coding-Expert
174
0

Themis-GGUF

license:apache-2.0
174
0

gpt2023-GGUF

NaNK
license:mit
174
0

Josiefied-Qwen2.5-7B-Instruct-abliterated-v2-GGUF

NaNK
license:apache-2.0
173
1

pythia-160m-deduped-GGUF

license:apache-2.0
173
0

llama2-exams-orca-sharegpt-GGUF

base_model:HWERI/llama2-exams-orca-sharegpt
172
0

DeepSeek-Coder-V2-Lite-Instruct-GGUF

171
0

AetherResearch_Cerebrum-1.0-7b-GGUF

NaNK
license:apache-2.0
171
0

phi-4-GGUF

NaNK
license:mit
170
1

Qwen1.5-32B-Chat-GGUF

NaNK
170
0

blossom-v3_1-yi-34b-GGUF

NaNK
license:apache-2.0
169
0

gemma-3-1b-it-GGUF

NaNK
169
0

BSC-LT_salamandraTA-2B-GGUF

NaNK
license:apache-2.0
168
0

deepseek-coder-6.7b-base-GGUF

NaNK
168
0

SuperNova-Medius-GGUF

license:apache-2.0
168
0

Llama-3.3-70B-Instruct-GGUF

NaNK
llama
167
9

mlx-community_Qwen3-4B-bf16-GGUF

NaNK
license:apache-2.0
167
0

Qwen2.5-Coder-1.5B-Instruct-GGUF

NaNK
license:apache-2.0
167
0

em_german_leo_mistral-GGUF

license:apache-2.0
167
0

google_gemma-3-1b-it-GGUF

NaNK
167
0

Phigments12-GGUF

NaNK
license:apache-2.0
166
0

SmolLM2-360M-Instruct-FT-GGUF

license:apache-2.0
166
0

DeepSeek-R1-Distill-Qwen-14B-GGUF

NaNK
165
1

Quyen-SE-v0.1-GGUF

NaNK
165
0

Llama-3.2-3B-GGUF

NaNK
llama
164
0

Quble_Test_Model_v1_Pretrain-GGUF

163
0

Mixtral-tiny-GGUF

163
0

cyber-risk-llama-3-8b-GGUF

NaNK
base_model:Vanessasml/cyber-risk-llama-3-8b
162
0

OpenLLM-France_Lucie-7B-GGUF

NaNK
llama-3
162
0

Viking-13B-GGUF

NaNK
license:apache-2.0
162
0

QuyXuan_documents-master-3B-GGUF

[](https://tensorblock.co) [](https://twitter.com/tensorblockaoi) [](https://discord.gg/Ej5NmeHFf2) [](https://github.com/TensorBlock) [](https://t.me/TensorBlock) Join our Discord to learn more about what we're building ↗ This repo contains GGUF format model files for QuyXuan/documents-master-3B. The files were quantized using machines provided by TensorBlock, and they are compatible with llama.cpp as of commit b5753. A comprehensive collection of Model Context Protocol (MCP) servers. A lightweight, open, and extensible multi-LLM interaction studio. | Filename | Quant type | File Size | Description | | -------- | ---------- | --------- | ----------- | | documents-master-3B-Q2K.gguf | Q2K | 1.364 GB | smallest, significant quality loss - not recommended for most purposes | | documents-master-3B-Q3KS.gguf | Q3KS | 1.543 GB | very small, high quality loss | | documents-master-3B-Q3KM.gguf | Q3KM | 1.687 GB | very small, high quality loss | | documents-master-3B-Q3KL.gguf | Q3KL | 1.815 GB | small, substantial quality loss | | documents-master-3B-Q40.gguf | Q40 | 1.917 GB | legacy; small, very high quality loss - prefer using Q3KM | | documents-master-3B-Q4KS.gguf | Q4KS | 1.928 GB | small, greater quality loss | | documents-master-3B-Q4KM.gguf | Q4KM | 2.019 GB | medium, balanced quality - recommended | | documents-master-3B-Q50.gguf | Q50 | 2.270 GB | legacy; medium, balanced quality - prefer using Q4KM | | documents-master-3B-Q5KS.gguf | Q5KS | 2.270 GB | large, low quality loss - recommended | | documents-master-3B-Q5KM.gguf | Q5KM | 2.322 GB | large, very low quality loss - recommended | | documents-master-3B-Q6K.gguf | Q6K | 2.644 GB | very large, extremely low quality loss | | documents-master-3B-Q80.gguf | Q80 | 3.422 GB | very large, extremely low quality loss - not recommended | Then, downoad the individual model file the a local directory If you wanna download multiple model files with a pattern (e.g., `Q4Kgguf`), you can try:

NaNK
llama
161
0

calme-3.1-qwenloi-3b-GGUF

NaNK
161
0

gemma-2-2b-it-abliterated-GGUF

NaNK
161
0

Qwen1.5-MoE-A2.7B-Chat-GGUF

NaNK
160
2

Meta-Llama-3-8B-hf-GGUF

NaNK
llama
160
0

Blue-Orchid-2x7b-GGUF

NaNK
license:apache-2.0
159
1

SeaLLM-7B-v2.5-GGUF

NaNK
159
0

Arcee-Spark-GGUF

license:apache-2.0
159
0

Velara-11B-V2-GGUF

NaNK
llama-2
159
0

ghost-8b-beta-GGUF

NaNK
llama
158
1

occiglot-7b-it-en-instruct-GGUF

NaNK
license:apache-2.0
158
0

Llama-3.1-Swallow-8B-Instruct-v0.3-GGUF

NaNK
base_model:tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3
157
2

anime-anything-promptgen-v2-GGUF

NaNK
157
2

Qwen1.5-1.8B-GGUF

NaNK
157
0

Llama-3.2-1B-GGUF

NaNK
llama
156
1

Goekdeniz-Guelmez_Josiefied-DeepSeek-R1-0528-Qwen3-8B-abliterated-v1-GGUF

NaNK
156
1

s1-0.5B-GGUF

NaNK
license:apache-2.0
156
1

Qwen2.5-Coder-7B-GGUF

NaNK
license:apache-2.0
156
0

Josiefied-Qwen2.5-14B-Instruct-abliterated-v4-GGUF

NaNK
license:apache-2.0
156
0

llama-3-2-1b-sft-GGUF

NaNK
license:llama3.2
156
0

TenyxChat-7B-v1-GGUF

NaNK
license:apache-2.0
156
0

gpt2-medium-GGUF

license:mit
155
0

agentica-org_DeepSWE-Preview-GGUF

license:mit
154
0

shuttleai_shuttle-3.5-GGUF

NaNK
license:apache-2.0
154
0

Meltemi-7B-Instruct-v1.5-GGUF

NaNK
license:apache-2.0
153
1

futurehouse_ether0-GGUF

NaNK
license:apache-2.0
153
0

Ambari-7B-Instruct-v0.1-sharded-GGUF

NaNK
license:mit
153
0

FreedomIntelligence_HuatuoGPT-Vision-7B-Qwen2.5VL-GGUF

NaNK
license:apache-2.0
153
0

Qwen1.5-32B-GGUF

NaNK
153
0

MiniCPM-2B-128k-GGUF

NaNK
152
1

gemma2-gutenberg-27B-GGUF

NaNK
152
1

Qwen2.5-Coder-32B-Instruct-GGUF

NaNK
license:apache-2.0
152
0

ChimeraLlama-3-8B-v2-GGUF

NaNK
base_model:mlabonne/ChimeraLlama-3-8B-v2
152
0

tnayajv2.0-GGUF

NaNK
license:apache-2.0
152
0

starcoder2-3b-GGUF

NaNK
151
2

saiga_llama3_8b-GGUF

NaNK
base_model:IlyaGusev/saiga_llama3_8b
151
1

SauerkrautLM-Qwen-32b-GGUF

NaNK
151
0

vinallama-2.7b-chat-GGUF

NaNK
base_model:vilm/vinallama-2.7b-chat
151
0

orca_mini_3b-GGUF

NaNK
license:cc-by-nc-sa-4.0
150
0

TarhanE_GRPO-Qwen3-0.6B-GGUF

[](https://tensorblock.co) [](https://twitter.com/tensorblockaoi) [](https://discord.gg/Ej5NmeHFf2) [](https://github.com/TensorBlock) [](https://t.me/TensorBlock) Join our Discord to learn more about what we're building ↗ This repo contains GGUF format model files for TarhanE/GRPO-Qwen3-0.6B. The files were quantized using machines provided by TensorBlock, and they are compatible with llama.cpp as of commit b5753. A comprehensive collection of Model Context Protocol (MCP) servers. A lightweight, open, and extensible multi-LLM interaction studio. | Filename | Quant type | File Size | Description | | -------- | ---------- | --------- | ----------- | | GRPO-Qwen3-0.6B-Q2K.gguf | Q2K | 0.296 GB | smallest, significant quality loss - not recommended for most purposes | | GRPO-Qwen3-0.6B-Q3KS.gguf | Q3KS | 0.323 GB | very small, high quality loss | | GRPO-Qwen3-0.6B-Q3KM.gguf | Q3KM | 0.347 GB | very small, high quality loss | | GRPO-Qwen3-0.6B-Q3KL.gguf | Q3KL | 0.368 GB | small, substantial quality loss | | GRPO-Qwen3-0.6B-Q40.gguf | Q40 | 0.382 GB | legacy; small, very high quality loss - prefer using Q3KM | | GRPO-Qwen3-0.6B-Q4KS.gguf | Q4KS | 0.383 GB | small, greater quality loss | | GRPO-Qwen3-0.6B-Q4KM.gguf | Q4KM | 0.397 GB | medium, balanced quality - recommended | | GRPO-Qwen3-0.6B-Q50.gguf | Q50 | 0.437 GB | legacy; medium, balanced quality - prefer using Q4KM | | GRPO-Qwen3-0.6B-Q5KS.gguf | Q5KS | 0.437 GB | large, low quality loss - recommended | | GRPO-Qwen3-0.6B-Q5KM.gguf | Q5KM | 0.444 GB | large, very low quality loss - recommended | | GRPO-Qwen3-0.6B-Q6K.gguf | Q6K | 0.495 GB | very large, extremely low quality loss | | GRPO-Qwen3-0.6B-Q80.gguf | Q80 | 0.639 GB | very large, extremely low quality loss - not recommended | Then, downoad the individual model file the a local directory If you wanna download multiple model files with a pattern (e.g., `Q4Kgguf`), you can try:

NaNK
150
0

Machina_24B.V2-GGUF

NaNK
149
1

reka-flash-3-GGUF

NaNK
license:apache-2.0
149
0

L3.2-Rogue-Creative-Instruct-Uncensored-Abliterated-7B-GGUF

NaNK
149
0

Unbabel_Tower-Plus-9B-GGUF

[](https://tensorblock.co) [](https://twitter.com/tensorblockaoi) [](https://discord.gg/Ej5NmeHFf2) [](https://github.com/TensorBlock) [](https://t.me/TensorBlock) Join our Discord to learn more about what we're building ↗ This repo contains GGUF format model files for Unbabel/Tower-Plus-9B. The files were quantized using machines provided by TensorBlock, and they are compatible with llama.cpp as of commit b5753. A comprehensive collection of Model Context Protocol (MCP) servers. A lightweight, open, and extensible multi-LLM interaction studio. | Filename | Quant type | File Size | Description | | -------- | ---------- | --------- | ----------- | | Tower-Plus-9B-Q2K.gguf | Q2K | 3.805 GB | smallest, significant quality loss - not recommended for most purposes | | Tower-Plus-9B-Q3KS.gguf | Q3KS | 4.338 GB | very small, high quality loss | | Tower-Plus-9B-Q3KM.gguf | Q3KM | 4.762 GB | very small, high quality loss | | Tower-Plus-9B-Q3KL.gguf | Q3KL | 5.132 GB | small, substantial quality loss | | Tower-Plus-9B-Q40.gguf | Q40 | 5.443 GB | legacy; small, very high quality loss - prefer using Q3KM | | Tower-Plus-9B-Q4KS.gguf | Q4KS | 5.479 GB | small, greater quality loss | | Tower-Plus-9B-Q4KM.gguf | Q4KM | 5.761 GB | medium, balanced quality - recommended | | Tower-Plus-9B-Q50.gguf | Q50 | 6.484 GB | legacy; medium, balanced quality - prefer using Q4KM | | Tower-Plus-9B-Q5KS.gguf | Q5KS | 6.484 GB | large, low quality loss - recommended | | Tower-Plus-9B-Q5KM.gguf | Q5KM | 6.647 GB | large, very low quality loss - recommended | | Tower-Plus-9B-Q6K.gguf | Q6K | 7.589 GB | very large, extremely low quality loss | | Tower-Plus-9B-Q80.gguf | Q80 | 9.827 GB | very large, extremely low quality loss - not recommended | Then, downoad the individual model file the a local directory If you wanna download multiple model files with a pattern (e.g., `Q4Kgguf`), you can try:

NaNK
license:cc-by-nc-sa-4.0
149
0

Viking-33B-GGUF

NaNK
license:apache-2.0
148
0

MBZUAI-Paris_Nile-Chat-4B-GGUF

NaNK
147
1

Breeze-7B-Instruct-v1_0-GGUF

NaNK
license:apache-2.0
147
0

Light-R1-32B-DS-GGUF

NaNK
license:apache-2.0
146
1

OpenHermes-2.5-Mistral-7B-pruned50-GGUF

NaNK
146
0

DeepSeek-R1-Distill-Qwen-Coder-32B-Fusion-9010-GGUF

NaNK
145
2

cognitivecomputations_samantha-mistral-instruct-7b-GGUF

NaNK
license:apache-2.0
145
1

mt0-xxl-mt-GGUF

NaNK
license:apache-2.0
145
0

swordfaith_ReTool-Qwen3-4B-SFT-cold-started-GGUF

[](https://tensorblock.co) [](https://twitter.com/tensorblockaoi) [](https://discord.gg/Ej5NmeHFf2) [](https://github.com/TensorBlock) [](https://t.me/TensorBlock) Join our Discord to learn more about what we're building ↗ This repo contains GGUF format model files for swordfaith/ReTool-Qwen3-4B-SFT-cold-started. The files were quantized using machines provided by TensorBlock, and they are compatible with llama.cpp as of commit b5753. A comprehensive collection of Model Context Protocol (MCP) servers. A lightweight, open, and extensible multi-LLM interaction studio. | Filename | Quant type | File Size | Description | | -------- | ---------- | --------- | ----------- | | ReTool-Qwen3-4B-SFT-cold-started-Q2K.gguf | Q2K | 1.669 GB | smallest, significant quality loss - not recommended for most purposes | | ReTool-Qwen3-4B-SFT-cold-started-Q3KS.gguf | Q3KS | 1.887 GB | very small, high quality loss | | ReTool-Qwen3-4B-SFT-cold-started-Q3KM.gguf | Q3KM | 2.076 GB | very small, high quality loss | | ReTool-Qwen3-4B-SFT-cold-started-Q3KL.gguf | Q3KL | 2.240 GB | small, substantial quality loss | | ReTool-Qwen3-4B-SFT-cold-started-Q40.gguf | Q40 | 2.370 GB | legacy; small, very high quality loss - prefer using Q3KM | | ReTool-Qwen3-4B-SFT-cold-started-Q4KS.gguf | Q4KS | 2.383 GB | small, greater quality loss | | ReTool-Qwen3-4B-SFT-cold-started-Q4KM.gguf | Q4KM | 2.497 GB | medium, balanced quality - recommended | | ReTool-Qwen3-4B-SFT-cold-started-Q50.gguf | Q50 | 2.824 GB | legacy; medium, balanced quality - prefer using Q4KM | | ReTool-Qwen3-4B-SFT-cold-started-Q5KS.gguf | Q5KS | 2.824 GB | large, low quality loss - recommended | | ReTool-Qwen3-4B-SFT-cold-started-Q5KM.gguf | Q5KM | 2.890 GB | large, very low quality loss - recommended | | ReTool-Qwen3-4B-SFT-cold-started-Q6K.gguf | Q6K | 3.306 GB | very large, extremely low quality loss | | ReTool-Qwen3-4B-SFT-cold-started-Q80.gguf | Q80 | 4.280 GB | very large, extremely low quality loss - not recommended | Then, downoad the individual model file the a local directory If you wanna download multiple model files with a pattern (e.g., `Q4Kgguf`), you can try:

NaNK
license:apache-2.0
145
0

Mistral_solar-slerp-GGUF

NaNK
license:apache-2.0
145
0

Uncensored_llama_3.2_3b_safetensors-GGUF

NaNK
base_model:ICEPVP8977/Uncensored_llama_3.2_3b_safetensors
144
1

aya-expanse-8b-GGUF

NaNK
license:cc-by-nc-4.0
143
3

Noromaid-v0.4-Mixtral-Instruct-8x7b-Zloss-GGUF

NaNK
license:cc-by-nc-4.0
143
1

X-ALMA-13B-Pretrain-GGUF

NaNK
license:mit
143
1

tempesthenno-nuslerp-0124-GGUF

NaNK
license:apache-2.0
143
0

zephyr-7b-beta-GGUF

NaNK
license:mit
143
0

deepseek-coder-7b-base-v1.5-GGUF

NaNK
143
0

ghost-7b-v0.9.0-GGUF

NaNK
license:mit
143
0

MARS-GGUF

license:llama3
143
0

MNCJihunKim_Mistral-7B-SlimOrca-OP-8k-GGUF

NaNK
142
0

MediKAI-GGUF

142
0

llama3.1_1B_adapted-GGUF

NaNK
142
0

Jarvis1111_DoctorAgent-RL-SFT-1k-Thinking-GGUF

license:apache-2.0
141
1

Qwen2-VL-2B-Instruct-GGUF

NaNK
license:apache-2.0
141
0

FractalAIResearch_Fathom-R1-14B-GGUF

NaNK
license:mit
141
0

castorini_rank_vicuna_7b_v1_fp16-GGUF

NaNK
license:llama2
141
0

llama3-diverce-ver1.0-GGUF

NaNK
base_model:sel303/llama3-diverce-ver1.0
141
0

Llama-3.1-8B-Lexi-Uncensored-GGUF

NaNK
base_model:Orenguteng/Llama-3.1-8B-Lexi-Uncensored
140
0

HuatuoGPT-o1-72B-GGUF

NaNK
license:apache-2.0
140
0

LLaMA3-iterative-DPO-final-ExPO-GGUF

base_model:chujiezheng/LLaMA3-iterative-DPO-final-ExPO
140
0

llama3_generative_qa_2-GGUF

NaNK
base_model:kitopang/llama3_generative_qa_2
139
1

Qwen_Qwen3-8B-MLX-bf16-GGUF

NaNK
license:apache-2.0
139
0

kakaocorp_kanana-1.5-2.1b-instruct-2505-GGUF

NaNK
license:apache-2.0
139
0

llama3-math-trans-sft-GGUF

139
0

Qwen_Qwen3-1.7B-GGUF

NaNK
license:apache-2.0
138
1

Hermes-3-Llama-3.1-70B-GGUF

NaNK
Llama-3
138
1

Llama-3-WhiteRabbitNeo-8B-v2.0-GGUF

NaNK
base_model:WhiteRabbitNeo/Llama-3-WhiteRabbitNeo-8B-v2.0
138
0

Smaug-34B-v0.1-GGUF

NaNK
license:apache-2.0
138
0

Sailor2-20B-Chat-GGUF

NaNK
license:apache-2.0
138
0

CodeLlama-70b-Python-hf-GGUF

NaNK
llama-2
137
0

TowerBase-7B-v0.1-GGUF

NaNK
license:cc-by-nc-4.0
137
0

Qwen2.5-Coder-32B-GGUF

NaNK
license:apache-2.0
137
0

TheBloke_Wizard-Vicuna-30B-Uncensored-fp16-GGUF

NaNK
137
0

WizardLM-30B-Uncensored-Guanaco-SuperCOT-30b-GGUF

NaNK
136
4

mixtralnt-4x7b-test-GGUF

NaNK
license:cc-by-nc-4.0
136
0

RankingGPT-bloom-560m-GGUF

license:mit
136
0

dolphin-2.9.1-llama-3-8b-GGUF

NaNK
base_model:dphn/dolphin-2.9.1-llama-3-8b
135
1

Llama-3-ELYZA-JP-8B-GGUF

NaNK
base_model:elyza/Llama-3-ELYZA-JP-8B
135
1

granite-7b-instruct-GGUF

NaNK
license:apache-2.0
135
0

granite-8b-code-instruct-4k-GGUF

NaNK
dataset:bigcode/commitpackft
135
0

cybersentinal-2.0-GGUF

NaNK
llama
135
0

CohereLabs_c4ai-command-r-08-2024-GGUF

license:cc-by-nc-4.0
134
1

c4ai-command-r-v01-GGUF

NaNK
license:cc-by-nc-4.0
134
1

mlabonne_gemma-3-12b-it-qat-abliterated-GGUF

[](https://tensorblock.co) [](https://twitter.com/tensorblockaoi) [](https://discord.gg/Ej5NmeHFf2) [](https://github.com/TensorBlock) [](https://t.me/TensorBlock) Join our Discord to learn more about what we're building ↗ This repo contains GGUF format model files for mlabonne/gemma-3-12b-it-qat-abliterated. The files were quantized using machines provided by TensorBlock, and they are compatible with llama.cpp as of commit b5753. A comprehensive collection of Model Context Protocol (MCP) servers. A lightweight, open, and extensible multi-LLM interaction studio. | Filename | Quant type | File Size | Description | | -------- | ---------- | --------- | ----------- | | gemma-3-12b-it-qat-abliterated-Q2K.gguf | Q2K | 4.768 GB | smallest, significant quality loss - not recommended for most purposes | | gemma-3-12b-it-qat-abliterated-Q3KS.gguf | Q3KS | 5.458 GB | very small, high quality loss | | gemma-3-12b-it-qat-abliterated-Q3KM.gguf | Q3KM | 6.009 GB | very small, high quality loss | | gemma-3-12b-it-qat-abliterated-Q3KL.gguf | Q3KL | 6.480 GB | small, substantial quality loss | | gemma-3-12b-it-qat-abliterated-Q40.gguf | Q40 | 6.887 GB | legacy; small, very high quality loss - prefer using Q3KM | | gemma-3-12b-it-qat-abliterated-Q4KS.gguf | Q4KS | 6.935 GB | small, greater quality loss | | gemma-3-12b-it-qat-abliterated-Q4KM.gguf | Q4KM | 7.301 GB | medium, balanced quality - recommended | | gemma-3-12b-it-qat-abliterated-Q50.gguf | Q50 | 8.232 GB | legacy; medium, balanced quality - prefer using Q4KM | | gemma-3-12b-it-qat-abliterated-Q5KS.gguf | Q5KS | 8.232 GB | large, low quality loss - recommended | | gemma-3-12b-it-qat-abliterated-Q5KM.gguf | Q5KM | 8.445 GB | large, very low quality loss - recommended | | gemma-3-12b-it-qat-abliterated-Q6K.gguf | Q6K | 9.661 GB | very large, extremely low quality loss | | gemma-3-12b-it-qat-abliterated-Q80.gguf | Q80 | 12.510 GB | very large, extremely low quality loss - not recommended | Then, downoad the individual model file the a local directory If you wanna download multiple model files with a pattern (e.g., `Q4Kgguf`), you can try:

NaNK
134
0

llama3-8B-DarkIdol-2.2-Uncensored-1048K-GGUF

NaNK
llama3
133
4

stablelm-3b-4e1t-GGUF

NaNK
license:cc-by-sa-4.0
133
0

marin-community_marin-8b-base-GGUF

NaNK
license:apache-2.0
133
0

JSL-MedLlama-3-8B-v1.0-GGUF

NaNK
llama-3-8b
133
0

Qwen2.5-1.5B-Instruct-GGUF

NaNK
license:apache-2.0
133
0

Qwen2.5-Coder-14B-Instruct-GGUF

NaNK
license:apache-2.0
133
0

granite-3.0-8b-instruct-GGUF

NaNK
license:apache-2.0
133
0

Qwen1.5-MoE-A2.7B-GGUF

NaNK
132
2

Salesforce_xgen-small-9B-instruct-r-GGUF

NaNK
license:cc-by-nc-4.0
132
0

salamandra-2b-instruct-GGUF

NaNK
license:apache-2.0
131
1

SOLAR-10.7B-v1.0-GGUF

NaNK
license:apache-2.0
131
0

redrix_patricide-12B-Unslop-Mell-GGUF

NaNK
license:apache-2.0
131
0

JSL-MedLlama-3-8B-v2.0-GGUF

NaNK
llama-3-8b
131
0

SSH_355M-GGUF

license:apache-2.0
131
0

TinyLlama-1.1B-intermediate-step-1195k-token-2.5T-GGUF

NaNK
base_model:TinyLlama/TinyLlama-1.1B-intermediate-step-1195k-token-2.5T
131
0

Yi-Ko-6B-Instruct-v1.0-GGUF

NaNK
license:cc-by-nc-4.0
131
0

codellama-13b-instruct-nf4-fp16-upscaled-GGUF

NaNK
base_model:arnavgrg/codellama-13b-instruct-nf4-fp16-upscaled
131
0

L3H10M-0000-GGUF

NaNK
license:apache-2.0
131
0

MLlamav1-GGUF

NaNK
base_model:Haesteining/MLlamav1
131
0

ONS-SOLAR-10.7B-v1.2-GGUF

NaNK
license:cc-by-nc-4.0
130
0

open-llama-3b-v2-elmv3-GGUF

NaNK
openllama
130
0

Llama-3.2-3B-Instruct-GGUF

NaNK
llama
130
0

llama-3-8b-GGUF

NaNK
llama
130
0

WiroAI-Finance-Qwen-1.5B-GGUF

NaNK
license:apache-2.0
130
0

sft-ds-140k-GGUF

license:llama3
130
0

saiga_tlite_8b-GGUF

NaNK
130
0

Amal-70b-GGUF

NaNK
130
0

seeklhy_codes-7b-spider-GGUF

NaNK
130
0

r1-1776-GGUF

license:mit
129
8

Qwen2.5-14B-Instruct-GGUF

NaNK
license:apache-2.0
129
1

Meta-Llama-3.1-8B-Instruct-GGUF

NaNK
llama
129
0

andresnowak_Qwen3-0.6B-instruction-finetuned-GGUF

NaNK
129
0

Samantha2.0-Phi4-ita-16bit-GGUF

NaNK
license:apache-2.0
128
0

bloom-1b1-GGUF

NaNK
128
0

llama2-13B-eugeneparkthebest-GGUF

NaNK
llama2-13B
128
0

Felladrin_TinyMistral-248M-Chat-v2-GGUF

NaNK
license:apache-2.0
128
0

L3.1-Suze-Vume-2-calc-GGUF

128
0

s1K_32b-GGUF

NaNK
llama-factory
128
0

Qwen_Qwen3-8B-GGUF

NaNK
license:apache-2.0
127
2

unsloth_Qwen3-30B-A3B-Instruct-2507-GGUF

NaNK
license:apache-2.0
127
1

Qwen2.5-7B-nerd-uncensored-v1.0-GGUF

NaNK
license:apache-2.0
127
1

Qwen_Qwen3-4B-GGUF

NaNK
license:apache-2.0
127
0

redpajama-3b-chat-GGUF

NaNK
license:cc-by-nc-2.0
127
0

CodeLlama-34b-Python-hf-GGUF

NaNK
llama-2
127
0

llama-3-8b-Instruct-GGUF

NaNK
base_model:AI-Sweden-Models/Llama-3-8B-instruct
127
0

LLaMA3-SFT-v2-GGUF

NaNK
base_model:RLHFlow/LLaMA3-SFT-v2
127
0

Llama-3-8B-Stroganoff-GGUF

NaNK
llama-3
127
0

GritLM-8x7B-GGUF

NaNK
license:apache-2.0
127
0

MiniPLM-Qwen-200M-GGUF

NaNK
license:apache-2.0
126
1

Mistral-7B-Merge-14-v0.3-GGUF

NaNK
license:apache-2.0
126
0

Orion-14B-Base-GGUF

NaNK
126
0

FusionNet-GGUF

license:mit
126
0

falcon-7b-instruct-GGUF

NaNK
license:apache-2.0
126
0

Aspik101_Vicuzard-30B-Uncensored-instruct-PL-lora_unload-GGUF

NaNK
llama
125
1

PhysicsWallahAI_Aryabhata-1.0-GGUF

[](https://tensorblock.co) [](https://twitter.com/tensorblockaoi) [](https://discord.gg/Ej5NmeHFf2) [](https://github.com/TensorBlock) [](https://t.me/TensorBlock) Join our Discord to learn more about what we're building ↗ This repo contains GGUF format model files for PhysicsWallahAI/Aryabhata-1.0. The files were quantized using machines provided by TensorBlock, and they are compatible with llama.cpp as of commit b5753. A comprehensive collection of Model Context Protocol (MCP) servers. A lightweight, open, and extensible multi-LLM interaction studio. | Filename | Quant type | File Size | Description | | -------- | ---------- | --------- | ----------- | | Aryabhata-1.0-Q2K.gguf | Q2K | 3.016 GB | smallest, significant quality loss - not recommended for most purposes | | Aryabhata-1.0-Q3KS.gguf | Q3KS | 3.492 GB | very small, high quality loss | | Aryabhata-1.0-Q3KM.gguf | Q3KM | 3.808 GB | very small, high quality loss | | Aryabhata-1.0-Q3KL.gguf | Q3KL | 4.088 GB | small, substantial quality loss | | Aryabhata-1.0-Q40.gguf | Q40 | 4.431 GB | legacy; small, very high quality loss - prefer using Q3KM | | Aryabhata-1.0-Q4KS.gguf | Q4KS | 4.458 GB | small, greater quality loss | | Aryabhata-1.0-Q4KM.gguf | Q4KM | 4.683 GB | medium, balanced quality - recommended | | Aryabhata-1.0-Q50.gguf | Q50 | 5.315 GB | legacy; medium, balanced quality - prefer using Q4KM | | Aryabhata-1.0-Q5KS.gguf | Q5KS | 5.315 GB | large, low quality loss - recommended | | Aryabhata-1.0-Q5KM.gguf | Q5KM | 5.445 GB | large, very low quality loss - recommended | | Aryabhata-1.0-Q6K.gguf | Q6K | 6.254 GB | very large, extremely low quality loss | | Aryabhata-1.0-Q80.gguf | Q80 | 8.099 GB | very large, extremely low quality loss - not recommended | Then, downoad the individual model file the a local directory If you wanna download multiple model files with a pattern (e.g., `Q4Kgguf`), you can try:

NaNK
license:cc-by-nc-4.0
125
0

CarrotAI_Rabbit-Ko-15B-Instruct-GGUF

NaNK
license:apache-2.0
125
0

LLaMAntino-3-ANITA-8B-Inst-DPO-ITA-GGUF

NaNK
llama
125
0

Samantha2.0-Phi3-Medium-ita-16bit-GGUF

NaNK
license:apache-2.0
125
0

granite-3.1-1b-a400m-base-GGUF

NaNK
license:apache-2.0
125
0

llama2-22b-chat-wizard-uncensored-GGUF

NaNK
base_model:nkpz/llama2-22b-chat-wizard-uncensored
125
0

L3-8B-Celeste-V1.2-GGUF

NaNK
llama-factory
125
0

DarkIdol-Llama-3.1-8B-Instruct-1.1-Uncensored-GGUF

NaNK
llama3
124
2

Mixtral-8x7B-v0.1-GGUF

NaNK
license:apache-2.0
124
0

StableBeluga-7B-GGUF

NaNK
124
0

cymist-2-v02-SFT-GGUF

license:apache-2.0
124
0

AceGPT-v2-32B-GGUF

NaNK
license:apache-2.0
124
0

Samantha-Qwen-2-7B-GGUF

NaNK
license:apache-2.0
124
0

trinity-v1-GGUF

NaNK
license:apache-2.0
124
0

Llama-3.1-Distilled-GGUF

base_model:DeepMount00/Llama-3.1-Distilled
124
0

DPO_model-GGUF

124
0

qwen1.5-0.5b-sft-GGUF

NaNK
124
0