ZeroAgency

56 models • 1 total models in database
Sort by:

Zero-Mistral-24B-gguf

NaNK
llama.cpp
432
4

Mistral-Small-3.2-24B-Instruct-2506-Text-Only

Modified Small 3.2: - No vision encoder - Standard "Mistral" architecture - Based on anthracite-core/Mistral-Small-3.2-24B-Instruct-2506-Text-Only - System prompt from 3.2 version added as default to chat template

NaNK
91
1

Zero-Mistral-24B

NaNK
license:mit
53
7

gpt-oss-20b-multilingual-reasoner-lora

This model is a fine-tuned version of openai/gpt-oss-20b. It has been trained using TRL. - PEFT 0.17.0 - TRL: 0.21.0 - Transformers: 4.55.0 - Pytorch: 2.8.0+cu128 - Datasets: 4.0.0 - Tokenizers: 0.21.4

NaNK
46
1

zero-mistral-beta62-e1

NaNK
32
0

zero-mistral-beta62-e2

NaNK
30
0

gpt-oss-20b-multilingual-reasoning

This model is a fine-tuned version of axolotl-ai-co/gpt-oss-20b-dequantized on the HuggingFaceH4/Multilingual-Thinking dataset. The following hyperparameters were used during training: - learningrate: 2e-05 - trainbatchsize: 4 - evalbatchsize: 4 - seed: 42 - distributedtype: multi-GPU - numdevices: 8 - totaltrainbatchsize: 32 - totalevalbatchsize: 32 - optimizer: Use OptimizerNames.ADAMWTORCHFUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizerargs=No additional optimizer arguments - lrschedulertype: constantwithwarmup - trainingsteps: 8 - Transformers 4.55.0 - Pytorch 2.8.0+cu128 - Datasets 4.0.0 - Tokenizers 0.21.4

NaNK
license:apache-2.0
23
0

Zero Mistral Small 24B Instruct 2501

NaNK
license:mit
21
5

gpt-oss-120b-multilingual-reasoning

This model is a fine-tuned version of axolotl-ai-co/gpt-oss-120b-dequantized on the HuggingFaceH4/Multilingual-Thinking dataset. The following hyperparameters were used during training: - learningrate: 2e-05 - trainbatchsize: 4 - evalbatchsize: 4 - seed: 42 - distributedtype: multi-GPU - numdevices: 8 - totaltrainbatchsize: 32 - totalevalbatchsize: 32 - optimizer: Use OptimizerNames.ADAMWTORCHFUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizerargs=No additional optimizer arguments - lrschedulertype: constantwithwarmup - trainingsteps: 8 - Transformers 4.55.0 - Pytorch 2.8.0+cu128 - Datasets 4.0.0 - Tokenizers 0.21.4

NaNK
license:apache-2.0
13
2

Mistral-Small-3.2-24B-Instruct-2506

NaNK
license:apache-2.0
9
1

Mistral-Small-3.1-24B-Instruct-2503-hf

NaNK
license:apache-2.0
5
2

Zero-Mistral-Small-24B-Instruct-2501-lora

NaNK
license:mit
5
0

Zero-Mistral-Small-3.1-24B-Instruct-2503-beta

NaNK
license:mit
5
0

zero-judge

NaNK
license:mit
5
0

gemma-3-4b-it-think-untrained

NaNK
5
0

o1_t-lite-it-1.0_lora

NaNK
license:mit
4
4

Zero-Mistral-Small-24B-Instruct-2501-Q4_K_M

NaNK
license:mit
4
0

zero-mistral-beta52-2

4
0

zero-mistral-beta57

4
0

zero-mistral-beta58

4
0

Zero-Gemma-12b-beta1

NaNK
3
1

Zero-Mistral-Small-24B-Instruct-2501-BF16

NaNK
license:mit
3
0

Zero-Mistral-Small-3.1-24B-Instruct-2503-beta6-GGUF

NaNK
3
0

zero-mistral-beta50-e1

3
0

zero-mistral-beta50-e2

3
0

zero-mistral-beta51-lora-e2

NaNK
3
0

zero-mistral-beta51-lora-e3

NaNK
3
0

Zero-Gemma-12b-beta2

NaNK
3
0

Zero-Gemma-3-12b-brs

NaNK
3
0

zero-mistral-beta54

3
0

zero-mistral-beta55

3
0

zero-mistral-beta60

3
0

zero-mistral-beta60-dpo7

3
0

Zero-Mistral-Small-24B-Instruct-2501-Q8_0

NaNK
license:mit
2
0

Zero-Mistral-Small-24B-Instruct-2501-F16

NaNK
license:mit
2
0

zero-mistral-beta50-e2.2

NaNK
2
0

zero-mistral-beta52-2.5e-5-1ks

2
0

zero-mistral-beta53-e2

NaNK
2
0

zero-mistral-beta57-e2

NaNK
2
0

zero-mistral-beta61

2
0

zero-mistral-beta60-e2

- Developed by: [More Information Needed] - Funded by [optional]: [More Information Needed] - Shared by [optional]: [More Information Needed] - Model type: [More Information Needed] - Language(s) (NLP): [More Information Needed] - License: [More Information Needed] - Finetuned from model [optional]: [More Information Needed] - Repository: [More Information Needed] - Paper [optional]: [More Information Needed] - Demo [optional]: [More Information Needed] Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: [More Information Needed] - Hours used: [More Information Needed] - Cloud Provider: [More Information Needed] - Compute Region: [More Information Needed] - Carbon Emitted: [More Information Needed]

NaNK
2
0

zero-summary-v2-beta15

This model is a fine-tuned version of ZeroAgency/zero-llama-3.1-8b-beta6 on the bethrezen/thinking-summary-v2 dataset. The following hyperparameters were used during training: - learningrate: 4e-05 - trainbatchsize: 1 - evalbatchsize: 1 - seed: 42 - distributedtype: multi-GPU - numdevices: 8 - totaltrainbatchsize: 8 - totalevalbatchsize: 8 - optimizer: Use OptimizerNames.ADAMWTORCHFUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizerargs=No additional optimizer arguments - lrschedulertype: cosine - lrschedulerwarmupsteps: 15 - numepochs: 2.0 - Transformers 4.49.0 - Pytorch 2.5.1+cu124 - Datasets 3.2.0 - Tokenizers 0.21.0

NaNK
llama
1
1

Zero-Mistral-Small-3.1-24B-Instruct-2503-beta6

NaNK
1
0

Zero-Gemma-12b-bs16-e1

NaNK
1
0

Zero-Gemma-12b-bs16-e2

NaNK
1
0

Zero-Gemma-12b-beta2-e1

NaNK
1
0

Zero-Gemma-3-12b-brs-e2

- Developed by: [More Information Needed] - Funded by [optional]: [More Information Needed] - Shared by [optional]: [More Information Needed] - Model type: [More Information Needed] - Language(s) (NLP): [More Information Needed] - License: [More Information Needed] - Finetuned from model [optional]: [More Information Needed] - Repository: [More Information Needed] - Paper [optional]: [More Information Needed] - Demo [optional]: [More Information Needed] Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: [More Information Needed] - Hours used: [More Information Needed] - Cloud Provider: [More Information Needed] - Compute Region: [More Information Needed] - Carbon Emitted: [More Information Needed]

NaNK
1
0

zero-mistral-beta52-2.5e-5-1ks-e2

NaNK
1
0

zero-mistral-beta53

1
0

zero-mistral-beta56-e2

- Developed by: [More Information Needed] - Funded by [optional]: [More Information Needed] - Shared by [optional]: [More Information Needed] - Model type: [More Information Needed] - Language(s) (NLP): [More Information Needed] - License: [More Information Needed] - Finetuned from model [optional]: [More Information Needed] - Repository: [More Information Needed] - Paper [optional]: [More Information Needed] - Demo [optional]: [More Information Needed] Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: [More Information Needed] - Hours used: [More Information Needed] - Cloud Provider: [More Information Needed] - Compute Region: [More Information Needed] - Carbon Emitted: [More Information Needed]

NaNK
1
0

zero-mistral-beta56

1
0

zero-mistral-beta57-e1

NaNK
1
0

zero-mistral-beta59

1
0

zero-mistral-beta60-dpo8

1
0

zero-llama-3.1-8b-beta6

NaNK
llama
0
1

zero-gemma-3-4b-it-beta4-e3

- Developed by: [More Information Needed] - Funded by [optional]: [More Information Needed] - Shared by [optional]: [More Information Needed] - Model type: [More Information Needed] - Language(s) (NLP): [More Information Needed] - License: [More Information Needed] - Finetuned from model [optional]: [More Information Needed] - Repository: [More Information Needed] - Paper [optional]: [More Information Needed] - Demo [optional]: [More Information Needed] Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: [More Information Needed] - Hours used: [More Information Needed] - Cloud Provider: [More Information Needed] - Compute Region: [More Information Needed] - Carbon Emitted: [More Information Needed]

NaNK
0
1