cstr
cohere-transcribe-03-2026-GGUF
Spaetzle-v60-7b
aya-expanse-8b-Q4_K_M-GGUF
cstr/aya-expanse-8b-Q4KM-GGUF This model was converted to GGUF format from `CohereForAI/aya-expanse-8b` using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model. Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well. Step 2: Move into the llama.cpp folder and build it with `LLAMACURL=1` flag along with other hardware-specific flags (for ex: LLAMACUDA=1 for Nvidia GPUs on Linux).
octen-embedding-0.6b-onnx-int4
granite-speech-4.0-1b-GGUF
Octen-Embedding-0.6B-ONNX-INT8-FULL
Llama3-DiscoLeo-Instruct-8B-v0.1-GGUF
Ministral-8B-Instruct-2410-GGUF
salamandra-7b-instruct-GGUF
whisper-large-v3-turbo-int8_float32
whisper-large-v3-turbo-german-int8_float32
Spaetzle-v85-7b-GGUF
Phi-3-mini-4k-instruct-LLaMAfied-GGUF
ALMA-7B-R-GGUF
Llama3-DiscoLeo-Instruct-8B-32k-v0.1-GGUF
DiscoLM_German_7b_v1_chat-GGUF
Llama3_DiscoLM_German_8b_v0.1_experimental-GGUF
llama3.1-8b-spaetzle-v74-GGUF
mt0-large-Q4_K_M-GGUF
cstr/mt0-large-Q4KM-GGUF This model was converted to GGUF format from `bigscience/mt0-large` using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model. Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well. Step 2: Move into the llama.cpp folder and build it with `LLAMACURL=1` flag along with other hardware-specific flags (for ex: LLAMACUDA=1 for Nvidia GPUs on Linux).