ContextualAI

71 models • 2 total models in database
Sort by:

archangel_sft-kto_llama13b

NaNK
llama
644
3

ctxl-rerank-v2-instruct-multilingual-6b

NaNK
license:cc-by-nc-sa-4.0
299
14

tiny-random-MistralForCausalLM

281
0

ctxl-rerank-v2-instruct-multilingual-2b

NaNK
license:cc-by-nc-sa-4.0
269
1

ctxl-rerank-v2-instruct-multilingual-1b

NaNK
license:cc-by-nc-sa-4.0
230
10

Llama-200M

llama
83
0

ctxl-rerank-v2-instruct-multilingual-2b-nvfp4

NaNK
license:cc-by-nc-sa-4.0
74
0

ctxl-rerank-v2-instruct-multilingual-1b-nvfp4

NaNK
license:cc-by-nc-sa-4.0
73
0

LMUnit-llama3.1-70b

NaNK
llama
57
2

Contextual KTO Mistral PairRM

This repo contains the model and tokenizer checkpoints for: - model family mistralai/Mistral-7B-Instruct-v0.2 - optimized with the loss KTO - aligned using the snorkelai/Snorkel-Mistral-PairRM-DPO-Dataset - via 3 iterations of KTO on one epoch of each training partition, each previous iteration's model serving as the reference for the subsequent. [03/06/2024]: We are #2 on the (verified) Alpaca Eval 2.0 Leaderboard scoring 33.23! To prompt this model, ensure that the format is consistent with that of TuluV2. For example, a prompt should be formatted as follows, where ` ` corresponds to the human's role and ` ` corresponds to the LLM's role. The human should speak first: Note that a beginning-of-sequence (BOS) token is automatically added at tokenization time and does not have to be added by you. No end-of-sequence (EOS) token is added to the prompt. You may also use our tokenizer's `applychattemplate` if doing inference with `chatml` set or evaluating generations through non-local clients. For more info on KTO refer to our code repository or blog for more details on the methodology. If you found this work useful, feel free to cite our work:

license:apache-2.0
38
32

ctx-bird-reward-250121

NaNK
26
3

ctxl-rerank-v2-instruct-multilingual-6b-nvfp4

NaNK
license:cc-by-nc-sa-4.0
26
2

LMUnit-qwen2.5-72b

NaNK
24
5

archangel_dpo_pythia1-4b

NaNK
license:apache-2.0
12
0

archangel_sft_llama7b

NaNK
llama
11
1

archangel_dpo_pythia6-9b

NaNK
license:apache-2.0
11
0

archangel_kto_llama7b

NaNK
llama
9
1

archangel_sft_pythia6-9b

NaNK
license:apache-2.0
8
0

archangel_kto_pythia2-8b

NaNK
license:apache-2.0
8
0

archangel_ppo_llama13b

NaNK
llama
8
0

archangel_ppo_pythia12-0b

NaNK
license:apache-2.0
7
0

archangel_kto_llama13b

NaNK
llama
6
0

archangel_sft-dpo_llama30b

NaNK
llama
6
0

archangel_sft-slic_llama7b

NaNK
llama
5
0

archangel_sft-kto_llama30b

NaNK
llama
4
2

archangel_sft_pythia1-4b

NaNK
license:apache-2.0
4
0

archangel_slic_pythia6-9b

NaNK
license:apache-2.0
4
0

archangel_ppo_llama30b

NaNK
llama
4
0

archangel_csft_llama7b

NaNK
llama
4
0

archangel_slic_llama7b

NaNK
llama
3
1

archangel_dpo_pythia2-8b

NaNK
license:apache-2.0
3
0

archangel_dpo_pythia12-0b

NaNK
license:apache-2.0
3
0

archangel_ppo_llama7b

NaNK
llama
3
0

archangel_sft-kto_pythia12-0b

NaNK
license:apache-2.0
3
0

archangel_sft-csft_pythia1-4b

NaNK
license:apache-2.0
3
0

archangel_sft-csft_pythia2-8b

NaNK
license:apache-2.0
3
0

archangel_sft-csft_pythia6-9b

NaNK
license:apache-2.0
3
0

archangel_sft-csft_llama13b

NaNK
llama
3
0

archangel_sft-slic_llama13b

NaNK
llama
3
0

archangel_kto_llama30b

NaNK
llama
2
2

archangel_sft_pythia12-0b

NaNK
license:apache-2.0
2
0

archangel_slic_pythia1-4b

NaNK
license:apache-2.0
2
0

archangel_slic_pythia2-8b

NaNK
license:apache-2.0
2
0

archangel_slic_pythia12-0b

NaNK
license:apache-2.0
2
0

archangel_kto_pythia6-9b

NaNK
license:apache-2.0
2
0

archangel_sft-ppo_pythia1-4b

NaNK
license:apache-2.0
2
0

archangel_sft-ppo_pythia2-8b

NaNK
license:apache-2.0
2
0

archangel_sft-ppo_llama13b

NaNK
llama
2
0

archangel_sft-slic_pythia2-8b

NaNK
license:apache-2.0
2
0

archangel_csft_pythia12-0b

NaNK
license:apache-2.0
2
0

archangel_sft-csft_llama30b

NaNK
llama
2
0

archangel_csft_llama30b

NaNK
llama
2
0

archangel_sft_pythia2-8b

NaNK
license:apache-2.0
1
1

archangel_sft_llama13b

NaNK
llama
1
0

archangel_slic_llama30b

NaNK
llama
1
0

archangel_dpo_llama7b

NaNK
llama
1
0

archangel_dpo_llama30b

NaNK
llama
1
0

archangel_ppo_pythia6-9b

NaNK
license:apache-2.0
1
0

archangel_sft-dpo_pythia6-9b

NaNK
license:apache-2.0
1
0

archangel_sft-kto_pythia2-8b

NaNK
license:apache-2.0
1
0

archangel_sft-kto_llama7b

NaNK
llama
1
0

archangel_sft-ppo_pythia12-0b

NaNK
license:apache-2.0
1
0

archangel_sft-ppo_llama7b

NaNK
llama
1
0

archangel_csft_pythia1-4b

NaNK
license:apache-2.0
1
0

archangel_sft-csft_pythia12-0b

NaNK
license:apache-2.0
1
0

archangel_sft-slic_pythia12-0b

NaNK
license:apache-2.0
1
0

archangel_sft-csft_llama7b

NaNK
llama
1
0

zephyr_sft_dpo

1
0

Llama-3.1-8b-Instruct

NaNK
llama
1
0

archangel_csft_pythia2-8b

NaNK
license:apache-2.0
0
2

zephyr_sft_kto

0
1