monsterapi
gpt2_alpaca-lora
gpt2_124m_norobots
gemma-2b-lora-maths-orca-200k
Gptj-6b_alpaca-gpt4
llama2-7b-tiny-codes-code-generation
mistral_7b_DolphinCoder
codellama_7b_DolphinCoder
mistral_7b_WizardLMEvolInstruct70k
llama2_SQL_Answers_finetuned
llama7B_alpaca-lora
gemma-2-2b-hindi-translator
sd21_anime_finetuning
opt125M_alpaca
OpenPlatypus_Falcon_7b
Mistral-7B-v0.1-Dolly-15k
CodeAlpaca_LLAMA2_7B
sdxl_car_finetuning
OpenPlatypus_LLAMA2_7b
opt1.3B_codeinstruct
falcon_7b_DolphinCoder
llama2_7b_DolphinCoder
codellama7b_codealpaca20k
Llama-3_1-8B-Instruct-orca-ORPO
Model Used: meta-llama/Meta-Llama-3.1-8B-Instruct Dataset: Intel/orcadpopairs The Intel Orca dataset is a specialized version of the OpenOrca dataset, which includes ~1M GPT-4 completions and ~3.2M GPT-3.5 completions. This dataset is tabularized to align with the distributions in the ORCA paper and focuses on preference optimization by clearly indicating which responses are good and which are bad. It is primarily used in natural language processing for training and evaluation. This finetuning run was performed using MonsterAPI's LLM finetuner with ORPO (Optimized Response Preference Optimization) for enhancing preference optimization. - Completed in a total duration of 1 hour and 39 minutes for 1 epoch. - Costed `$2.69` for the entire process. - Epochs: 1 - Cost Per Epoch: $2.69 - Total Finetuning Cost: $2.69 - Model Path: meta-llama/Meta-Llama-3.1-8B-Instruct - Learning Rate: 0.001 - Data Split: 90% train 10% validation - Gradient Accumulation Steps: 16