monsterapi

50 models • 2 total models in database
Sort by:

gpt2_alpaca-lora

NaNK
32
1

gpt2_124m_norobots

NaNK
license:apache-2.0
22
1

gemma-2b-lora-maths-orca-200k

NaNK
license:apache-2.0
9
7

Gptj-6b_alpaca-gpt4

NaNK
8
0

llama2-7b-tiny-codes-code-generation

NaNK
llama2
7
1

mistral_7b_DolphinCoder

NaNK
license:apache-2.0
7
0

codellama_7b_DolphinCoder

NaNK
code-llama
6
1

mistral_7b_WizardLMEvolInstruct70k

NaNK
meta-llama
5
11

llama2_SQL_Answers_finetuned

NaNK
meta-llama/Llama-2-7b
5
3

llama7B_alpaca-lora

NaNK
llama1-7b
5
1

gemma-2-2b-hindi-translator

NaNK
license:apache-2.0
5
1

sd21_anime_finetuning

NaNK
5
0

opt125M_alpaca

4
0

OpenPlatypus_Falcon_7b

NaNK
base_model:codellama/CodeLlama-7b-hf
4
0

Mistral-7B-v0.1-Dolly-15k

NaNK
llama
4
0

CodeAlpaca_LLAMA2_7B

NaNK
llama2-7b
3
4

sdxl_car_finetuning

NaNK
3
3

OpenPlatypus_LLAMA2_7b

NaNK
meta-llama/Llama-2-7b-hf
3
1

opt1.3B_codeinstruct

NaNK
base_model:codellama/CodeLlama-7b-hf
3
0

falcon_7b_DolphinCoder

NaNK
license:apache-2.0
3
0

llama2_7b_DolphinCoder

NaNK
llama2
3
0

codellama7b_codealpaca20k

NaNK
codellama7b
2
2

Llama-3_1-8B-Instruct-orca-ORPO

Model Used: meta-llama/Meta-Llama-3.1-8B-Instruct Dataset: Intel/orcadpopairs The Intel Orca dataset is a specialized version of the OpenOrca dataset, which includes ~1M GPT-4 completions and ~3.2M GPT-3.5 completions. This dataset is tabularized to align with the distributions in the ORCA paper and focuses on preference optimization by clearly indicating which responses are good and which are bad. It is primarily used in natural language processing for training and evaluation. This finetuning run was performed using MonsterAPI's LLM finetuner with ORPO (Optimized Response Preference Optimization) for enhancing preference optimization. - Completed in a total duration of 1 hour and 39 minutes for 1 epoch. - Costed `$2.69` for the entire process. - Epochs: 1 - Cost Per Epoch: $2.69 - Total Finetuning Cost: $2.69 - Model Path: meta-llama/Meta-Llama-3.1-8B-Instruct - Learning Rate: 0.001 - Data Split: 90% train 10% validation - Gradient Accumulation Steps: 16

NaNK
llama
2
2

falcon-7b-python-code-instructions-18k-alpaca

NaNK
license:apache-2.0
2
1

sdxl_finetuning_anime

NaNK
2
1

sdxl_chinatown_finetuning

NaNK
2
1

Falcon_40B_dolly15k

NaNK
license:apache-2.0
2
0

gpt2

license:mit
2
0

CodeLlama-70b-hf_4bit_bnb

NaNK
llama
2
0

Llama3.3_70b

NaNK
llama
2
0

zephyr-7b-alpha_metamathqa

NaNK
license:apache-2.0
1
4

llama2_7b_WizardLMEvolInstruct70k

NaNK
meta-llama
1
2

Falcon_180B_dolly15k

NaNK
license:apache-2.0
1
1

llama2_7b_norobots

NaNK
llama2
1
1

mistral_7b_HalfEpoch_DolphinCoder

NaNK
license:apache-2.0
1
1

Mixtral-8x7B-v0.1_4bit_bnb

NaNK
1
1

llama2_70B_dolly15k

NaNK
meta-llama
1
0

gpt2_124m_WizardLMEvolInstruct70k

NaNK
license:apache-2.0
1
0

zephyr-7b-beta-CTranslate2-bfloat16

NaNK
license:apache-2.0
1
0

falcon_7b_OpenPlatypus

NaNK
license:apache-2.0
1
0

opt-350m_4bit_bnb

NaNK
1
0

falcon-40b_4bit_bnb

NaNK
license:apache-2.0
1
0

Meta-Llama-3-70B-Instruct_4bit_bnb

NaNK
llama
1
0

llama2-code-generation

NaNK
llama2
0
10

mistral_7b_norobots

NaNK
license:apache-2.0
0
4

falcon_7b_norobots

NaNK
license:apache-2.0
0
2

zephyr_7b_norobots

NaNK
license:apache-2.0
0
2

zephyr-7b-WizardLM-alpaca-instruct-70k-unfiltered

NaNK
license:apache-2.0
0
1

falcon_7b_3epoch_norobots

NaNK
license:apache-2.0
0
1

Mixtral-8x7B-Instruct-v0.1_4bit_bnb

NaNK
license:apache-2.0
0
1