danielus
12 models • 1 total models in database
Sort by:
Qwen3 Coder REAP 25B A3B Q4 K M GGUF
danielus/Qwen3-Coder-REAP-25B-A3B-Q4KM-GGUF This model was converted to GGUF format from `cerebras/Qwen3-Coder-REAP-25B-A3B` using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model. Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well. Step 2: Move into the llama.cpp folder and build it with `LLAMACURL=1` flag along with other hardware-specific flags (for ex: LLAMACUDA=1 for Nvidia GPUs on Linux).
NaNK
llama-cpp
324
2
Mermaid-Phi-3-medium-128k-instruct-14B-Q6_K-GGUF
NaNK
llama-cpp
15
0
MermaidSolar-Q6_K-GGUF
llama-cpp
14
0
MermaidSolar-Q5_K_M-GGUF
llama-cpp
12
0
Mermaid-Phi-3-medium-128k-instruct-14B-Q8_0-GGUF
NaNK
llama-cpp
9
0
Llama-3-8b-Ita-gguf
NaNK
license:mit
8
1
MermaidMixtral-3x7b-Q6_K-GGUF
NaNK
llama-cpp
6
0
MermaidMixtral-3x7b-Q8_0-GGUF
NaNK
llama-cpp
6
0
MermaidSolar-Q8_0-GGUF
llama-cpp
4
0
MermaidSolar-Q4_K_S-GGUF
llama-cpp
4
0
Mermaid-Llama-3-8B-Q8_0-GGUF
NaNK
llama-cpp
4
0
ggml-whisper-models
license:mit
0
4