Quazim0t0

30 models • 13 total models in database
Sort by:

Phi4.React.Turn.V2.Full

NaNK
llama
59
1

Phi4.React.Turn.v2

NaNK
llama
27
0

Phi4.Turn.R1Distill_v1.5.1_Q4_k-GGUF

NaNK
llama
25
5

Phi4.Turn.R1Distill_v1.5_Q4_k-GGUF

NaNK
llama
25
1

ThinkPhi.Turn1.1-q4_k_m-GGUF

NaNK
llama
23
0

model-gguf

- Developed by: Quazim0t0 - License: apache-2.0 - Finetuned from model : Quazim0t0/Amethyst-1 This qwen3 model was trained 2x faster with Unsloth and Huggingface's TRL library.

NaNK
license:apache-2.0
23
0

Phi4.Turn.R1Distill_v1.4_Q4_k-GGUF

NaNK
llama
19
1

Phi4.Turn.React.V2.SafeTensors

NaNK
llama
18
1

Phi4.Turn.R1Distill_v1.2_Q4_k

- Developed by: Quazim0t0 - License: apache-2.0 - Finetuned from model : unsloth/phi-4-unsloth-bnb-4bit - GGUF - Trained for 8 Hours on A800 with the Bespoke Stratos 17k Dataset. - Trained for 6 Hours on A800 with the Bespoke Stratos 35k Dataset. - 10$ Training...I'm actually amazed by the results. Model hasn't been tested yet, will update when model has been. If using this model for Open WebUI here is a simple function to organize the models responses: https://openwebui.com/f/quaz93/phi4turnr1distillthoughtfunctionv1 Overview These LoRA adapters were trained using diverse reasoning datasets that incorporate structured Thought and Solution responses to enhance logical inference. This project was designed to test the R1 dataset on Phi-4, aiming to create a lightweight, fast, and efficient reasoning model. All adapters were fine-tuned using an NVIDIA A800 GPU, ensuring high performance and compatibility for continued training, merging, or direct deployment. As part of an open-source initiative, all resources are made publicly available for unrestricted research and development. LoRA Adapters Below are the currently available LoRA fine-tuned adapters (as of January 30, 2025): - Phi4.Turn.R1Distill-Lora1 - Phi4.Turn.R1Distill-Lora2 - Phi4.Turn.R1Distill-Lora3 - Phi4.Turn.R1Distill-Lora4 - Phi4.Turn.R1Distill-Lora5 - Phi4.Turn.R1Distill-Lora6 - Phi4.Turn.R1Distill-Lora7 - Phi4.Turn.R1Distill-Lora8 GGUF Full & Quantized Models To facilitate broader testing and real-world inference, GGUF Full and Quantized versions have been provided for evaluation on Open WebUI and other LLM interfaces. Version 1 - Phi4.Turn.R1Distill.Q80 - Phi4.Turn.R1Distill.Q4k - Phi4.Turn.R1Distill.16bit Loading LoRA Adapters with `transformers` and `peft` To load and apply the LoRA adapters on Phi-4, use the following approach: ```python from transformers import AutoModelForCausalLM, AutoTokenizer from peft import PeftModel basemodel = "microsoft/Phi-4" loraadapter = "Quazim0t0/Phi4.Turn.R1Distill-Lora1" tokenizer = AutoTokenizer.frompretrained(basemodel) model = AutoModelForCausalLM.frompretrained(basemodel) model = PeftModel.frompretrained(model, loraadapter)

NaNK
llama
18
0

Qwen2.5_11

12
0

Amethyst-1-GGUF

NaNK
license:apache-2.0
9
0

Phi4.Turn.R1Distill_v1.5.1-Tensors

Language: English. License: Apache 2.0.

NaNK
llama
7
4

model

- Developed by: Quazim0t0 - License: apache-2.0 - Finetuned from model : Quazim0t0/Amethyst-1 This qwen3 model was trained 2x faster with Unsloth and Huggingface's TRL library.

NaNK
license:apache-2.0
7
0

ThinkPhi1.1-Tensors

Language: English. License: Apache 2.0.

NaNK
llama
6
2

Phi4.Turn.R1Distill.Q4_k

NaNK
llama
5
2

Phi4.Turn.R1Distill_v1.3_Q4_k-GGUF

NaNK
llama
5
1

Phi4.Turn.R1Distill_v1.1_Q4_k

NaNK
llama
4
0

Qwen2.5_11-Tens

NaNK
license:apache-2.0
4
0

Amethyst-1

NaNK
license:apache-2.0
3
0

ThinkPhiExperimental5-q4_k_m-GGUF

NaNK
llama
2
1

Phi4.Turn.R1Distill.Q8_0

- Developed by: Quazim0t0 - License: apache-2.0 - Finetuned from model : unsloth/phi-4-unsloth-bnb-4bit - GGUF - Trained for 8 Hours on A800 with the Bespoke Stratos 17k Dataset. - 10$ Training...I'm actually amazed by the results. If using this model for Open WebUI here is a simple function to organize the models responses: https://openwebui.com/f/quaz93/phi4turnr1distillthoughtfunctionv1 Overview These LoRA adapters were trained using diverse reasoning datasets that incorporate structured Thought and Solution responses to enhance logical inference. This project was designed to test the R1 dataset on Phi-4, aiming to create a lightweight, fast, and efficient reasoning model. All adapters were fine-tuned using an NVIDIA A800 GPU, ensuring high performance and compatibility for continued training, merging, or direct deployment. As part of an open-source initiative, all resources are made publicly available for unrestricted research and development. LoRA Adapters Below are the currently available LoRA fine-tuned adapters (as of January 30, 2025): - Phi4.Turn.R1Distill-Lora1 - Phi4.Turn.R1Distill-Lora2 - Phi4.Turn.R1Distill-Lora3 - Phi4.Turn.R1Distill-Lora4 - Phi4.Turn.R1Distill-Lora5 - Phi4.Turn.R1Distill-Lora6 - Phi4.Turn.R1Distill-Lora7 - Phi4.Turn.R1Distill-Lora8 GGUF Full & Quantized Models To facilitate broader testing and real-world inference, GGUF Full and Quantized versions have been provided for evaluation on Open WebUI and other LLM interfaces. Version 1 - Phi4.Turn.R1Distill.Q80 - Phi4.Turn.R1Distill.Q4k - Phi4.Turn.R1Distill.16bit Loading LoRA Adapters with `transformers` and `peft` To load and apply the LoRA adapters on Phi-4, use the following approach: ```python from transformers import AutoModelForCausalLM, AutoTokenizer from peft import PeftModel basemodel = "microsoft/Phi-4" loraadapter = "Quazim0t0/Phi4.Turn.R1Distill-Lora1" tokenizer = AutoTokenizer.frompretrained(basemodel) model = AutoModelForCausalLM.frompretrained(basemodel) model = PeftModel.frompretrained(model, loraadapter)

NaNK
llama
1
1

Amethyst-0

NaNK
license:apache-2.0
1
1

Rosemary-14b

Base model includes Quazim0t0 time 14b stock and Quazim0t0 Mithril 14B sce.

NaNK
llama
1
0

Fugazi14b-Q8_0-GGUF

NaNK
llama-cpp
1
0

Fugazi14b-Q4_K_M-GGUF

NaNK
llama-cpp
1
0

Phi4.Turn.R1Distill-v1.0-Tensors

NaNK
llama
0
1

Alien-GGUF

0
1

ThinkPhiExperimental6-q4_k_m-GGUF

NaNK
llama
0
1

ThinkPhiExperimental7-GGUF

- Developed by: Quazim0t0 - License: apache-2.0 - Finetuned from model : unsloth/phi-4-unsloth-bnb-4bit This llama model was trained 2x faster with Unsloth and Huggingface's TRL library.

NaNK
llama
0
1

Rune-14b

NaNK
llama
0
1