oobabooga
CodeBooga-34B-v0.1
1) Phind-CodeLlama-34B-v2 2) WizardCoder-Python-34B-V1.0 It was created with the BlockMerge Gradient script, the same one that was used to create MythoMax-L2-13b, and with the same settings. The following YAML was used: Both base models use the Alpaca format, so it should be used for this one as well. I made a quick experiment where I asked a set of 3 Python and 3 Javascript questions (real world, difficult questions with nuance) to the following models: 1) This one 2) A second variant generated with `modelpath1` and `modelpath2` swapped in the YAML above, which I called CodeBooga-Reversed-34B-v0.1 3) WizardCoder-Python-34B-V1.0 4) Phind-CodeLlama-34B-v2 Specifically, I used 4.250b EXL2 quantizations of each. I then sorted the responses for each question by quality, and attributed the following scores: 4th place: 0 3rd place: 1 2nd place: 2 1st place: 4 CodeBooga-34B-v0.1: 22 WizardCoder-Python-34B-V1.0: 12 Phind-CodeLlama-34B-v2: 7 CodeBooga-Reversed-34B-v0.1: 1 CodeBooga-34B-v0.1 performed very well, while its variant performed poorly, so I uploaded the former but not the latter. TheBloke has kindly provided GGUF quantizations for llama.cpp: