win10
MagKr-3.2-24B-thinking
GPT-OSS-26B-abliterated-Preview-Q4_K_M-GGUF
SVD-Qwen3-Coder-Next-Thinking
Pixtral-12B-2409-hf-text-only-Q8_0-GGUF
NeMoria-21b-Q8_0-GGUF
nemolita-21b-Q8_0-GGUF
35b-beta-long-Q4_K_M-GGUF
Mistral-Nemo-Instruct-2407-Q4_K_M-GGUF
RYS-Gemma-2-27b-it-Q6_K-GGUF
Phi-3.5-24-10-06-Q8_0-GGUF
phi-3.5-Sakura-Yuzu-v1.5-7.64b
aya-expanse-32b-Q5_K_M-GGUF
win10/aya-expanse-32b-Q5KM-GGUF This model was converted to GGUF format from `CohereForAI/aya-expanse-32b` using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model. Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well. Step 2: Move into the llama.cpp folder and build it with `LLAMACURL=1` flag along with other hardware-specific flags (for ex: LLAMACUDA=1 for Nvidia GPUs on Linux).
DeepSeek-Coder-V2-Lite-Instruct-Q6_K-GGUF
aya-expanse-32b-Q4_K_M-GGUF
win10/aya-expanse-32b-Q4KM-GGUF This model was converted to GGUF format from `CohereForAI/aya-expanse-32b` using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model. Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well. Step 2: Move into the llama.cpp folder and build it with `LLAMACURL=1` flag along with other hardware-specific flags (for ex: LLAMACUDA=1 for Nvidia GPUs on Linux).
GPT-OSS-26B-abliterated-Preview-Q8_0-GGUF
Mistral-Nemo-abliterated-Nemo-Pro-v2
DeepSeek-Coder-V2-Lite-Instruct-Q8_0-GGUF
Mistral-Nemo-Instruct-2407-20b-Q5_K_M-GGUF
phi-3.5-sakura-yuzu-v2-Q8_0-GGUF
Mistral-rp-24b-karcher-Q6_K-GGUF
GPT-OSS-26B-abliterated-Preview
This is an expanded version of unsloth/gpt-oss-20b-BF16 scaled up to 26B parameters and created with abliteration (see abliteration to know more about it). Usage You can use this model in your applications by loading it with Hugging Face's `transformers` library: - Risk of Sensitive or Controversial Outputs: This model's safety filtering has been significantly reduced, potentially generating sensitive, controversial, or inappropriate content. Users should exercise caution and rigorously review generated outputs. - Not Suitable for All Audiences: Due to limited content filtering, the model's outputs may be inappropriate for public settings, underage users, or applications requiring high security. - Legal and Ethical Responsibilities: Users must ensure their usage complies with local laws and ethical standards. Generated content may carry legal or ethical risks, and users are solely responsible for any consequences. - Research and Experimental Use: It is recommended to use this model for research, testing, or controlled environments, avoiding direct use in production or public-facing commercial applications. - Monitoring and Review Recommendations: Users are strongly advised to monitor model outputs in real-time and conduct manual reviews when necessary to prevent the dissemination of inappropriate content. - No Default Safety Guarantees: Unlike standard models, this model has not undergone rigorous safety optimization. The author bears no responsibility for any consequences arising from its use. Donation Your donation helps us continue our further development and improvement, a cup of coffee can do it. - PayPal: Support via PayPal - Ko-fi: Support our work on Ko-fi
phi3.5-mini-24-09-30-Q8_0-GGUF
Qwerky-QwQ-32B-Q5_K_M-GGUF
GPT-OSS-30B-Preview
nemolita-21b
openhands-Nemotron-32B-karcher-Q4_K_M-GGUF
win10/openhands-Nemotron-32B-karcher-Q4KM-GGUF This model was converted to GGUF format from `mergekit-community/openhands-Nemotron-32B-karcher` using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model. Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well. Step 2: Move into the llama.cpp folder and build it with `LLAMACURL=1` flag along with other hardware-specific flags (for ex: LLAMACUDA=1 for Nvidia GPUs on Linux).
Phi-3.5-24-10-06
phi-3.5-sakura-yuzu-v2.5-Q8_0-GGUF
Breeze-7B-FC-v1-0-EvolKit-75K-nopm_claude_writing_fixed-adapter-F32-GGUF
taide-meta-it-16b
phi-3.5-Sakura-Yuzu-Q8_0-GGUF
phi3.5-mini-24-09-30
Norns-Qwen2.5-7B-Q8_0-GGUF
phi-3.5-sakura-yuzu-v3.0-Q8_0-GGUF
Blue-Rose-Coder-12.3B-Instruct-Q8_0-GGUF
MagiDevs-24B-2506-Vision-Q8_0-GGUF
win10/MagiDevs-24B-2506-Vision-Q80-GGUF This model was converted to GGUF format from `win10/MagiDevs-24B-2506-Vision` using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model. Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well. Step 2: Move into the llama.cpp folder and build it with `LLAMACURL=1` flag along with other hardware-specific flags (for ex: LLAMACUDA=1 for Nvidia GPUs on Linux).
Fused-Yi-Qwen3-3B
ArliAI-RPMax-v1.3-merge-8B-Q8_0-GGUF
Meta-Llama-3-12B-Instruct-Q6_K-GGUF
Qwen2.5-5B-Instruct
phi-3.5-Sakura-Yuzu-v1.5-7.64b-Q8_0-GGUF
granite-3.0-8b-instruct-Q8_0-GGUF
Mistral-Nemo-abliterated-Nemo-Pro-v2-Q8_0-GGUF
Mistral-RP-24b-karcher-pro
llama3-13.45b-Instruct-Q6_K-GGUF
llama3-13.45b-Instruct-Q8_0-GGUF
llama3-13.45b-Instruct-Q5_K_M-GGUF
llama3-13.45b-Instruct-Q4_K_M-GGUF
Phi-3.5-mini-instruct-Q8_0-GGUF
phi-3.5-sakura-yuzu-v2
Weirdslerp2-25B-Q5_K_M-GGUF
Infinirc-ArliAI-RPMax-v1.3-merge-13.3B-Q8_0-GGUF
win10/Infinirc-ArliAI-RPMax-v1.3-merge-13.3B-Q80-GGUF This model was converted to GGUF format from `win10/Infinirc-ArliAI-RPMax-v1.3-merge-13.3B` using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model. Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well. Step 2: Move into the llama.cpp folder and build it with `LLAMACURL=1` flag along with other hardware-specific flags (for ex: LLAMACUDA=1 for Nvidia GPUs on Linux).
miscii-Virtuoso-Small-Q8_0-GGUF
MagiDevs-24B-2506-Vision
MagiDevs-24B-2506-Vision-Q6_K-GGUF
win10/MagiDevs-24B-2506-Vision-Q6K-GGUF This model was converted to GGUF format from `win10/MagiDevs-24B-2506-Vision` using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model. Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well. Step 2: Move into the llama.cpp folder and build it with `LLAMACURL=1` flag along with other hardware-specific flags (for ex: LLAMACUDA=1 for Nvidia GPUs on Linux).
phi3-128k-6b
DarkIdol-Llama-3.1-13.3B-Instruct-1.2-Uncensored
miscii-14b-1028-Q8_0-GGUF
Norns-Qwen2.5-12B
This is a merge of pre-trained language models created using mergekit. This model was merged using the passthrough merge method. The following models were included in the merge: win10/Norns-Qwen2.5-7B The following YAML configuration was used to produce this model:
SphinxMind-14B-normalize-false-Q8_0-GGUF
EVA-Instruct-QwQ-32B-Preview-Q4_K_M-GGUF
win10/EVA-Instruct-QwQ-32B-Preview-Q4KM-GGUF This model was converted to GGUF format from `win10/EVA-Instruct-QwQ-32B-Preview` using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model. Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well. Step 2: Move into the llama.cpp folder and build it with `LLAMACURL=1` flag along with other hardware-specific flags (for ex: LLAMACUDA=1 for Nvidia GPUs on Linux).
Breeze-13B-32k-Instruct-v1_0-Q8_0-GGUF
internlm2_5-20b-chat-abliterated-Q6_K-GGUF
mergekit-karcher-pifptpx-Q8_0-GGUF
Mistral-v0.3-13B-32k-Base-v1
Qwen2.5-mini-Instruct-2
Infinirc-ArliAI-RPMax-v1.3-merge-8B-Q8_0-GGUF
Norns-Qwen2.5-7B-v0.2-Q8_0-GGUF
shuttle-3-mini-Q8_0-GGUF
win10/shuttle-3-mini-Q80-GGUF This model was converted to GGUF format from `shuttleai/shuttle-3-mini` using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model. Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well. Step 2: Move into the llama.cpp folder and build it with `LLAMACURL=1` flag along with other hardware-specific flags (for ex: LLAMACUDA=1 for Nvidia GPUs on Linux).
Mistral-RP-24b-karcher-pro-Q4_K_M-GGUF
win10/Mistral-RP-24b-karcher-pro-Q4KM-GGUF This model was converted to GGUF format from `win10/Mistral-RP-24b-karcher-pro` using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model. Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well. Step 2: Move into the llama.cpp folder and build it with `LLAMACURL=1` flag along with other hardware-specific flags (for ex: LLAMACUDA=1 for Nvidia GPUs on Linux).
magnum-v2-12b-Q8_0-GGUF
InternLM2_5-20B-ArliAI-RPMax-v1.1-Q6_K-GGUF
ArliAI-RPMax-v1.3-merge-llama3-8B-Q8_0-GGUF
ChatML-Nemo-Pro-Q8_0-GGUF
sthenno-Test-maybe-is-pro-v2-Q8_0-GGUF
Lingshu-32B-Q4_K_M-GGUF
win10/Lingshu-32B-Q4KM-GGUF This model was converted to GGUF format from `lingshu-medical-mllm/Lingshu-32B` using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model. Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well. Step 2: Move into the llama.cpp folder and build it with `LLAMACURL=1` flag along with other hardware-specific flags (for ex: LLAMACUDA=1 for Nvidia GPUs on Linux).
EVA-QwQ-32B-Preview
Please support my ko-fi. https://ko-fi.com/ogodwin10 my-paypal: https://www.paypal.com/ncp/payment/X7DMN9DUBH2X8 If you like this model, please give me some sponsorship and I will continue to create better merges. If there is a device that can fine-tune the model, there will be more new models fine-tuned after the merge (just like solar 10b). merge This is a merge of pre-trained language models created using mergekit. This model was merged using the TIES merge method using Qwen/QwQ-32B-Preview as a base. The following models were included in the merge: EVA-UNIT-01/EVA-Qwen2.5-32B-v0.2 The following YAML configuration was used to produce this model:
DeepSeek-R1-Distill-sthenno-14b-0121
Your support = more models My Ko-fi page (Click here) This is a merge of pre-trained language models created using mergekit. This model was merged using the TIES merge method using sthenno-com/miscii-14b-1225 as a base. The following models were included in the merge: deepseek-ai/DeepSeek-R1-Distill-Qwen-14B sthenno/tempesthenno-ppo-ckpt40 The following YAML configuration was used to produce this model:
Phitis-14b-Base
karcher-test-32b
SphinxMind-14B-Q8_0-GGUF
win10/SphinxMind-14B-Q80-GGUF This model was converted to GGUF format from `win10/SphinxMind-14B` using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model. Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well. Step 2: Move into the llama.cpp folder and build it with `LLAMACURL=1` flag along with other hardware-specific flags (for ex: LLAMACUDA=1 for Nvidia GPUs on Linux).
Lusca-33B-Q4_K_M-GGUF
Llama-3.2-3B-Instruct-24-9-29
phi-3.5-Sakura-Yuzu-v1.5
Mistral-Nemo-Instruct-2407-20b-Q8_0-GGUF
Mistral-Nemo-Instruct-2407-20b-Q4_K_M-GGUF
ArliAI-RPMax-v1.3-merge-13.3B-Q8_0-GGUF
win10/ArliAI-RPMax-v1.3-merge-13.3B-Q80-GGUF This model was converted to GGUF format from `win10/ArliAI-RPMax-v1.3-merge-13.3B` using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model. Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well. Step 2: Move into the llama.cpp folder and build it with `LLAMACURL=1` flag along with other hardware-specific flags (for ex: LLAMACUDA=1 for Nvidia GPUs on Linux).
Norns-Qwen2.5-Coder-7B-v0.1-Q8_0-GGUF
high-speed-mixing-7B-V1-Q8_0-GGUF
EVA-Meissa-mini-pro-v2-Q8_0-GGUF
win10/EVA-Meissa-mini-pro-v2-Q80-GGUF This model was converted to GGUF format from `win10/EVA-Meissa-mini-pro-v2` using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model. Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well. Step 2: Move into the llama.cpp folder and build it with `LLAMACURL=1` flag along with other hardware-specific flags (for ex: LLAMACUDA=1 for Nvidia GPUs on Linux).
openhands-Nemotron-32B-karcher-300-Q4_K_M-GGUF
win10/openhands-Nemotron-32B-karcher-300-Q4KM-GGUF This model was converted to GGUF format from `mergekit-community/openhands-Nemotron-32B-karcher-300` using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model. Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well. Step 2: Move into the llama.cpp folder and build it with `LLAMACURL=1` flag along with other hardware-specific flags (for ex: LLAMACUDA=1 for Nvidia GPUs on Linux).
miscii-14b-1M-0128
Qwen2.5-2B-Instruct
Qwen2.5-2B-Instruct is a merge of the following models using LazyMergekit: Qwen/Qwen2.5-1.5B-Instruct Qwen/Qwen2.5-1.5B-Instruct Qwen/Qwen2.5-1.5B-Instruct Qwen/Qwen2.5-1.5B-Instruct Qwen/Qwen2.5-1.5B-Instruct Qwen/Qwen2.5-1.5B-Instruct Qwen/Qwen2.5-1.5B-Instruct Qwen/Qwen2.5-1.5B-Instruct Qwen/Qwen2.5-1.5B-Instruct Qwen/Qwen2.5-1.5B-Instruct Qwen/Qwen2.5-1.5B-Instruct Qwen/Qwen2.5-1.5B-Instruct Qwen/Qwen2.5-1.5B-Instruct Qwen/Qwen2.5-1.5B-Instruct Qwen/Qwen2.5-1.5B-Instruct Qwen/Qwen2.5-1.5B-Instruct Qwen/Qwen2.5-1.5B-Instruct Qwen/Qwen2.5-1.5B-Instruct Qwen/Qwen2.5-1.5B-Instruct Qwen/Qwen2.5-1.5B-Instruct Qwen/Qwen2.5-1.5B-Instruct Qwen/Qwen2.5-1.5B-Instruct Qwen/Qwen2.5-1.5B-Instruct Qwen/Qwen2.5-1.5B-Instruct Qwen/Qwen2.5-1.5B-Instruct Qwen/Qwen2.5-1.5B-Instruct
steiner-32b-preview-Q4_K_M-GGUF
WhiteRabbitNeo-2.5-Qwen-2.5-Coder-12.3B
Blue-Rose-Coder-12.3B-Instruct
Norns-Qwen2.5-12B-Q8_0-GGUF
win10/Norns-Qwen2.5-12B-Q80-GGUF This model was converted to GGUF format from `win10/Norns-Qwen2.5-12B` using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model. Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well. Step 2: Move into the llama.cpp folder and build it with `LLAMACURL=1` flag along with other hardware-specific flags (for ex: LLAMACUDA=1 for Nvidia GPUs on Linux).
EVA-Norns-Qwen2.5-v0.1-Q8_0-GGUF
win10/EVA-Norns-Qwen2.5-v0.1-Q80-GGUF This model was converted to GGUF format from `win10/EVA-Norns-Qwen2.5-v0.1` using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model. Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well. Step 2: Move into the llama.cpp folder and build it with `LLAMACURL=1` flag along with other hardware-specific flags (for ex: LLAMACUDA=1 for Nvidia GPUs on Linux).
miscii-14b-1225-Q8_0-GGUF
Qwen1.5-0.5b-Xia-Ai
phi-3.5-Sakura-Yuzu
phi-3.5-Sakura-Yuzu-v1.5-Q8_0-GGUF
DeepSeek-V2-Lite-XiaAi-Q8_0-GGUF
Qwen2.5-Math-12.3B-Instruct
This is a merge of pre-trained language models created using mergekit. This model was merged using the passthrough merge method. The following models were included in the merge: Qwen/Qwen2.5-Math-7B-Instruct The following YAML configuration was used to produce this model:
WhiteRabbitNeo-2.5-Qwen-2.5-Coder-12.3B-Q8_0-GGUF
ArliAI-RPMax-v1.3-merge-13.3B
This is a merge of pre-trained language models created using mergekit. This model was merged using the passthrough merge method. The following models were included in the merge: win10/ArliAI-RPMax-v1.3-merge-8B The following YAML configuration was used to produce this model:
Urdandi-Qwen2.5-7B
Norns-Qwen2.5-Coder-7B-Instruct-v0.1-Q8_0-GGUF
falcon-mamba-7b-instruct-Q8_0-GGUF
ChatML-Nemo-Pro-model_stock-Q8_0-GGUF
tempesthenno-ppo-ckpt40-Q8_0-GGUF
MagiD-24B
Llama-3.2-3B-F1-Instruct-vectormemory
EVA-QwQ-32B-Coder-Preview
my kofi: https://ko-fi.com/ogodwin10 If you like this model, please give me some sponsorship and I will continue to create better merges. If there is a device that can fine-tune the model, there will be more new models fine-tuned after the merge (just like solar 10b). merge This is a merge of pre-trained language models created using mergekit. This model was merged using the TIES merge method using EVA-UNIT-01/EVA-Qwen2.5-32B-v0.2 as a base. The following models were included in the merge: Qwen/QwQ-32B-Preview Qwen/Qwen2.5-Coder-32B-Instruct The following YAML configuration was used to produce this model:
Verdandi-Qwen2.5-7B
high-speed-mixing-7B-V2-Q8_0-GGUF
EVA-QwQ-32B-Preview-Q4_K_M-GGUF
my kofi: https://ko-fi.com/ogodwin10 If you like this model, please give me some sponsorship and I will continue to create better merges. If there is a device that can fine-tune the model, there will be more new models fine-tuned after the merge (just like solar 10b). win10/EVA-QwQ-32B-Preview-Q4KM-GGUF This model was converted to GGUF format from `win10/EVA-QwQ-32B-Preview` using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model. Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well. Step 2: Move into the llama.cpp folder and build it with `LLAMACURL=1` flag along with other hardware-specific flags (for ex: LLAMACUDA=1 for Nvidia GPUs on Linux).
wizardcoder-33b-v1.1-mirror-Q2_K-GGUF
phi-3.5-sakura-yuzu-v2.5
Qwen2.5-Coder-12.3b-Instruct-Q8_0-GGUF
Norns-Qwen2.5-7B
Norns-Qwen2.5-Coder-7B-v0.1
EVA-Meissa-mini-pro
EVA-Meissa-mini-pro-Q8_0-GGUF
win10/EVA-Meissa-mini-pro-Q80-GGUF This model was converted to GGUF format from `win10/EVA-Meissa-mini-pro` using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model. Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well. Step 2: Move into the llama.cpp folder and build it with `LLAMACURL=1` flag along with other hardware-specific flags (for ex: LLAMACUDA=1 for Nvidia GPUs on Linux).
ChatML-Nemo-Pro-V2-Q8_0-GGUF
OuteTTS-500M-Fgo
karcher-max-iter1000-32b-Q4_K_M-GGUF
mergekit-karcher-pifptpx
KwaiCoder-AutoThink-preview-Q4_K_M-GGUF
win10/KwaiCoder-AutoThink-preview-Q4KM-GGUF This model was converted to GGUF format from `Kwaipilot/KwaiCoder-AutoThink-preview` using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model. Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well. Step 2: Move into the llama.cpp folder and build it with `LLAMACURL=1` flag along with other hardware-specific flags (for ex: LLAMACUDA=1 for Nvidia GPUs on Linux).
ERNIE-4.5-29B-A4B-PT
miscii-Virtuoso-Small
This is a merge of pre-trained language models created using mergekit. This model was merged using the TIES merge method using sthenno-com/miscii-14b-1028 as a base. The following models were included in the merge: arcee-ai/Virtuoso-Small The following YAML configuration was used to produce this model:
llama3-13.45b-Instruct
This is a merge of pre-trained language models created using mergekit. This model was merged using the passthrough merge method. The following models were included in the merge: F:\text-generation-webui\models\meta-llamaMeta-Llama-3-8B-Instruct The following YAML configuration was used to produce this model:
Qwen2.5-Coder-12.3b-Instruct
ChatML-Nemo-Pro-V2
Breeze-13B-32k-Base-v1_0
Breeze-13B-32k-Instruct-v1_0
Breeze-13B-32k-Instruct-v10 is a merge of the following models using mergekit: MediaTek-Research/Breeze-7B-32k-Instruct-v10 MediaTek-Research/Breeze-7B-32k-Instruct-v10 MediaTek-Research/Breeze-7B-32k-Instruct-v10 MediaTek-Research/Breeze-7B-32k-Instruct-v10 MediaTek-Research/Breeze-7B-32k-Instruct-v10 MediaTek-Research/Breeze-7B-32k-Instruct-v10 MediaTek-Research/Breeze-7B-32k-Instruct-v10
Llama-3.2-3B-Instruct-24-9-29-Q8_0-GGUF
MagpieLM-8B
dolphin-2.9.3-mistral-nemo-20b-V2
ghost-13.3b-beta-1608
Urd-Qwen2.5-7B
Norns-Qwen2.5-Coder-7B-Instruct-v0.1
ChatML-Nemo-Pro-weight-density-increase-test-Q8_0-GGUF
sthenno-Test-maybe-is-pro
karcher-max-iter1000-32b
yi-qwen3-16b
DeepSeek-R1-Distill-sthenno-14b-0121-union-tokenizer
EVA-Instruct-QwQ-32B-Preview
ArliAI-RPMax-v1.3-merge-8B
ChatML-Nemo-Pro
Nemotron2Gemma-AURORA-LoRA-27B-IT-0p95
Meta-Llama-3-15B-Instruct
Qwen2-12.3B
phi3.5-pro-10-08
Qwen2.5-mini-Instruct
DeepSeek-V2-Lite-XiaAi
This is a merge of pre-trained language models created using mergekit. This model was merged using the TIES merge method using E:\tabbyAPI\models\deepseek-aiDeepSeek-V2-Lite as a base. The following models were included in the merge: E:\tabbyAPI\models\deepseek-aiDeepSeek-Coder-V2-Lite-Instruct The following YAML configuration was used to produce this model:
Meissa-Qwen2.5-12.3B-Instruct
ArliAI-RPMax-v1.3-merge-llama3-8B
Infinirc-ArliAI-RPMax-v1.3-merge-8B
Norns-Qwen2.5-7B-v0.2
EVA-Norns-Qwen2.5-v0.1
This is a merge of pre-trained language models created using mergekit. This model was merged using the TIES merge method using EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1 as a base. The following models were included in the merge: win10/Urdandi-Qwen2.5-7B The following YAML configuration was used to produce this model: