sm54

21 models • 2 total models in database
Sort by:

GLM 4.6 MXFP4 MOE

NaNK
994
7

GLM-4.6-REAP-268B-A32B-128GB-GGUF

NaNK
592
1

GLM 4.6 REAP 218B A32B MXFP4 MOE

NaNK
580
3

GLM-4.6-REAP-268B-A32B-MXFP4_MOE

NaNK
315
1

Qwen3-235B-A22B-Thinking-2507-MXFP4_MOE

NaNK
285
1

Qwen3-235B-A22B-Instruct-2507-MXFP4_MOE

NaNK
118
9

Qwen3-235B-A22B-Thinking-2507-OPT-GGUF

NaNK
83
0

Qwen3-30B-A3B-Thinking-2507-OPT-GGUF

NaNK
73
0

Qwen3-Nemotron-32B-RLBFF-Q4_K_M

NaNK
59
0

GLM-4.5-MXFP4_MOE

NaNK
39
2

Qwen3-30B-A3B-Thinking-2507-Q5_K_M-GGUF

sm54/Qwen3-30B-A3B-Thinking-2507-Q5KM-GGUF This model was converted to GGUF format from `Qwen/Qwen3-30B-A3B-Thinking-2507` using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model. Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well. Step 2: Move into the llama.cpp folder and build it with `LLAMACURL=1` flag along with other hardware-specific flags (for ex: LLAMACUDA=1 for Nvidia GPUs on Linux).

NaNK
llama-cpp
33
1

QwQ-DeepSeek-R1-SkyT1-Flash-Lightest-32B-Q4_K_M-GGUF

NaNK
llama-cpp
18
2

gemma-3-27b-it-Q4_K_M-GGUF

NaNK
llama-cpp
18
0

OpenReasoning-Nemotron-14B-Q6_K-GGUF

sm54/OpenReasoning-Nemotron-14B-Q6K-GGUF This model was converted to GGUF format from `nvidia/OpenReasoning-Nemotron-14B` using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model. Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well. Step 2: Move into the llama.cpp folder and build it with `LLAMACURL=1` flag along with other hardware-specific flags (for ex: LLAMACUDA=1 for Nvidia GPUs on Linux).

NaNK
llama-cpp
11
0

OpenReasoning-Nemotron-32B-Q4_K_M-GGUF

sm54/OpenReasoning-Nemotron-32B-Q4KM-GGUF This model was converted to GGUF format from `nvidia/OpenReasoning-Nemotron-32B` using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model. Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well. Step 2: Move into the llama.cpp folder and build it with `LLAMACURL=1` flag along with other hardware-specific flags (for ex: LLAMACUDA=1 for Nvidia GPUs on Linux).

NaNK
llama-cpp
7
1

QwQ-DeepSeek-R1-SkyT1-Flash-Light-32B

NaNK
3
0

QwQ-DeepSeek-R1-SkyT1-Flash-Lighter-32B

NaNK
3
0

Absolute_Zero_Reasoner-Coder-14b-Q6_K-GGUF

NaNK
llama-cpp
3
0

QwQ-DeepSeek-R1-SkyT1-Flash-Lightest-32B

NaNK
2
4

FuseO1-QwQ-SkyT1-Flash-32B

NaNK
1
3

QwQ-DeepSeek-R1-SkyT1-Flash-Lighter-32B-Q4_K_M-GGUF

NaNK
llama-cpp
1
0