lemon07r

30 models • 17 total models in database
Sort by:

gemma-3-27b-it-antislop-Q4_K_M-GGUF

NaNK
llama-cpp
110
0

VellumMini-0.1-Qwen3-14B-Q4_K_M-GGUF

A quick static GGUF made with llama.cpp b6691 by converting HF safetensors to F32 gguf then quantizing down to q4km. I suggest using better quants, like imatrix quants if they're available. Model Page: https://huggingface.co/lemon07r/VellumMini-0.1-Qwen3-14B Just a sneak peek of what I'm cooking in a little project called Vellum. This model was made to evaluate the quality of the CreativeGPT dataset, and how well Qwen3 trains on it. This is just one of many datasets that the final model will be trained on (which will also be using a different base model). This got pretty good results compared to the regular instruct in my testing so thought I would share. I trained for 3 epochs, but both checkpoints at 2 epoch and 3 epoch were too overbaked. This checkpoint, at 1 epoch performed best. I'm pretty surprised how decent this came out since Qwen models aren't that great at writing, especially at this size. Use with thinking/chain-of-thought disabled. Use ChatML prompt format. Big thanks to everyone over at the KoboldAI discord. The members there have helped me a ton with various things over the long while I've been there. Training Details Parent Model https://huggingface.co/Qwen/Qwen3-14B Dataset(s) https://huggingface.co/datasets/N8Programs/CreativeGPT

NaNK
license:apache-2.0
21
1

Gemma-2-Ataraxy-9B

NaNK
20
77

Qwen3-R1-SLERP-DST-8B-Q4_K_S-Q8_0-GGUF

NaNK
license:apache-2.0
19
1

Qwen3-R1-SLERP-Q3T-8B-Q4_K_S-Q8_0-GGUF

NaNK
license:apache-2.0
18
3

Gemma-2-Ataraxy-Remix-9B-Q8_0-GGUF

NaNK
16
4

RiverCub-Gemma-3-27B

NaNK
11
3

Gemma-2-Ataraxy-v4d-9B

For all intents and purposes, we can consider this the best "all rounder" of the Ataraxy. While primarily made with creative writing in mind, this one has done well in testing, and was made based on a lot of what I've discovered through trial, experimentation, testing and feedback from others. Is this the best Ataraxy model? Not sure. I made a lot of variations, and quite honestly most of them aren't great, or at least as good as the very first version. The v2 series could do well in writing tests, but was a little too over the top and sloppy. The v3 series was a return to roots, and is a lot closer to v1, and can be considered v1 but slightly better or different and where we start to see some improvements in some areas. v4 is where we see further improvements, especially in overall or general use, even though my primary goal was writing ability. People seem to really like the very first version of Ataraxy, even if it doesn't do as well in various benchmarks. I hope this one comes close to beating it's predecessor, but if it doesn't I will keep trying. All the Ataraxy series are primarily made for writing ability, but after some threshold, it started to get hard to tell, and even test for writing performance because they were all pretty good. Hopefully with some feedback we can continue to seek improvements. GGUF Static: https://huggingface.co/mradermacher/Gemma-2-Ataraxy-v4d-9B-GGUF GGUF IMatrix: https://huggingface.co/mradermacher/Gemma-2-Ataraxy-v4d-9B-i1-GGUF This is a merge of pre-trained language models created using mergekit. This model was merged using the SLERP merge method. The following models were included in the merge: sam-paech/Darkest-muse-v1 lemon07r/Gemma-2-Ataraxy-v4c-9B The following YAML configuration was used to produce this model:

NaNK
9
19

RiverCub-Gemma-3-27B-Q4_K_S

This is a quick static GGUF quant of lemon07r/RiverCub-Gemma-3-27B in Q4KS. There will probably be better quants available eventually, from others. A slerp merge of what I believe to be the two best gemma 3 27b models after extensive testing of many. Unfortunately most finetunes of this model seem to make it come out worse than google's official instruct trained model, hence why I am using it in this slerp merge to keep some of it's magic. It really is quite good. Big tiger gemma v3 was surprisingly pretty good too, and seemed much less lobotomized compared to a lot of the other models I tested. Big thanks to everyone over at the KoboldAI discord. The members there have helped me a ton with various things over the long while I've been there, even letting me borrow GPU hours on runpod for some testing at some point. ɛmpti gets today's special thanks in particular for helping me figure out how to get rid of the extra head that was carried over from drummer's model.. which seems to have been caused by an issue with axolotl. This model was merged using the SLERP merge method. The following models were included in the merge: unsloth/gemma-3-27b-it TheDrummer/Big-Tiger-Gemma-27B-v3 The following YAML configuration was used to produce this model:

NaNK
9
2

Lllama-3-RedMagic4-8B-Q8_0-GGUF

NaNK
base_model:NousResearch/Meta-Llama-3-8B
9
0

VellumMini-0.1-Qwen3-14B

VellumMini-0.1-Qwen3-14B Just a sneak peek of what I'm cooking in a little project called Vellum. This model was made to evaluate the quality of the CreativeGPT dataset, and how well Qwen3 trains on it. This is just one of many datasets that the final model will be trained on (which will also be using a different base model). This got pretty good results compared to the regular instruct in my testing so thought I would share. I trained for 3 epochs, but both checkpoints at 2 epoch and 3 epoch were too overbaked. This checkpoint, at 1 epoch performed best. I'm pretty surprised how decent this came out since Qwen models aren't that great at writing, especially at this size. Use with thinking/chain-of-thought disabled. Use ChatML prompt format. - bartowski - https://huggingface.co/bartowski/lemon07rVellumMini-0.1-Qwen3-14B-GGUF - mradermacher - https://huggingface.co/mradermacher/VellumMini-0.1-Qwen3-14B-i1-GGUF - mradermacher - https://huggingface.co/mradermacher/VellumMini-0.1-Qwen3-14B-GGUF - Q4KM Only - https://huggingface.co/lemon07r/VellumMini-0.1-Qwen3-14B-Q4KM-GGUF Big thanks to everyone over at the KoboldAI discord. The members there have helped me a ton with various things over the long while I've been there. Training Details Parent Model https://huggingface.co/Qwen/Qwen3-14B Dataset(s) https://huggingface.co/datasets/N8Programs/CreativeGPT

NaNK
license:apache-2.0
6
2

Qwen3-R1-SLERP-Q3T-8B

NaNK
5
10

Gemma-2-Ataraxy-v4-Advanced-9B

NaNK
4
6

Gemma-2-Ataraxy-Remix-9B

NaNK
3
6

Gemma-2-Ataraxy-Advanced-9B

NaNK
3
5

Gemma-2-Ataraxy-v3i-9B-Q8_0-GGUF

NaNK
3
3

Gemma-2-Ataraxy-v2f-9B

NaNK
3
2

Gemma-2-Ataraxy-v3j-9B

Err another test model. We near the end of all this though. This is a merge of pre-trained language models created using mergekit. This model was merged using the della merge method using unsloth/gemma-2-9b-it as a base. The following models were included in the merge: ifable/gemma-2-Ifable-9B wzhouad/gemma-2-9b-it-WPO-HB nbeerbower/Gemma2-Gutenberg-Doppel-9B The following YAML configuration was used to produce this model:

NaNK
3
2

Qwen3-R1-SLERP-DST-8B

NaNK
3
1

Gemma-2-Ataraxy-v2-9B

NaNK
2
18

Gemma-2-Ataraxy-v4a-Advanced-9B

NaNK
2
5

Gemma-2-Ataraxy-v4c-9B

NaNK
2
5

Gemma-2-Ataraxy-Doppel-9B

NaNK
1
4

Gemma-2-Ataraxy-v3i-9B

NaNK
1
3

Gemma-2-Ataraxy-v3b-9B

This is a merge of pre-trained language models created using mergekit. This model was merged using the SLERP merge method. The following models were included in the merge: nbeerbower/Gemma2-Gutenberg-Doppel-9B wzhouad/gemma-2-9b-it-WPO-HB The following YAML configuration was used to produce this model:

NaNK
1
3

Gemma-2-Ataraxy-v3-Advanced-9B

This is a merge of pre-trained language models created using mergekit. This model was merged using the SLERP merge method. The following models were included in the merge: lemon07r/Gemma-2-Ataraxy-Advanced-9B nbeerbower/Gemma2-Gutenberg-Doppel-9B The following YAML configuration was used to produce this model:

NaNK
1
3

Gemma-2-Ataraxy-v2a-9B

Test/alternate version. Ignore for now. Or you can try it. May end up deleted. Will see what I want to do from here. This is a merge of pre-trained language models created using mergekit. This model was merged using the della merge method using unsloth/gemma-2-9b-it as a base. The following models were included in the merge: jsgreenawalt/gemma-2-9B-it-advanced-v2.1 lemon07r/Gemma-2-Ataraxy-9B ifable/gemma-2-Ifable-9B The following YAML configuration was used to produce this model:

NaNK
1
2

Gemma-2-Ataraxy-v4b-9B

NaNK
1
2

llama-3-NeuralMahou-8b

NaNK
llama
1
1

Llama-3-RedMagic4-8B

NaNK
llama
1
0