Rombo-Org
Qwen_QwQ-32B-GGUF_QX_k_Bf16
Rombo-LLM-V2.7-llama-3.2-1b
reka-flash-3-GGUF_QX_k_Bf16
Rombo-LLM-V3.0-Qwen-32b
Rombo-LLM-V2.5-Qwen-3b
Rombo-LLM-V2.5-Qwen-14b
Rombo-LLM-V2.7-gemma-2-27b
Rombo-LLM-V2.5-Qwen-7b
This model is based on the Qwen/Qwen2.5-7B-Instruct architecture and utilizes the transformers library.
Rombo-LLM-V2.7-gemma-2-9b
Rombo-LLM-V3.1-QWQ-32b_GGUF_Q8_0
Rombo-LLM-V2.7-mist-nemo-12b
Rombo-LLM-V2.7-llama-3.2-3b
Rombo-LLM-V2.7-gemma-2-27b_gguf
Rombo-LLM-V3.1-QWQ-32b
Rombo-LLM-V3.0-Qwen-72b
Rombo-LLM-V2.5-Qwen-0.5b
Rombo-LLM-V2.5-Qwen-1.5b
Rombo-LLM-V2.5-Qwen-32b
Rombos-LLM-V2.5-Qwen-32b is a continues finetuned version of Qwen2.5-32B. I noticed recently that the Qwen team did not learn from my methods of continuous finetuning, the great benefits, and no downsides of it. So I took it upon myself to merge the instruct model with the base model myself using the Ties merge method This version of the model shows higher performance than the original instruct and base models. GGUF: https://huggingface.co/bartowski/Replete-LLM-V2.5-Qwen-32b-GGUF - https://huggingface.co/Apel-sin/rombos-llm-v2.5-qwen-32b-exl2 - https://huggingface.co/async0x42/Rombos-LLM-V2.5-Qwen-32b-exl25.0bpw - https://huggingface.co/rombodawg/Rombos-LLM-V2.5-Qwen-32b-Exl2-4.25-bit