Yukang
Llama-2-7b-longlora-32k-ft
LongAlpaca-70B
LongAlpaca-7B
Llama-2-13b-chat-longlora-32k-sft
Llama-2-13b-longlora-64k
Llama-2-7b-longlora-100k-ft
Llama-2-70b-chat-longlora-32k-sft
Llama-2-13b-longlora-16k-ft
Llama-2-7b-longlora-16k-ft
LongAlpaca-13B
Llama 2 13b Longlora 32k Ft
LongAlpaca-13B-16k
LongAlpaca-70B-16k
Qwen2.5-3B-Open-R1-Code-GRPO
This model is a fine-tuned version of Qwen/Qwen2.5-3B-Instruct on the open-r1/verifiable-coding-problems-python dataset. It has been trained using TRL. This model was trained with GRPO, a method introduced in DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models. - TRL: 0.18.0 - Transformers: 4.52.3 - Pytorch: 2.6.0 - Datasets: 3.6.0 - Tokenizers: 0.21.1
Llama-2-7b-longlora-32k
Qwen2.5-7B-Open-R1-GRPO
This model is a fine-tuned version of Qwen/Qwen2.5-7B-Instruct on the open-r1/OpenR1-Math-220k dataset. It has been trained using TRL. This model was trained with GRPO, a method introduced in DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models. - TRL: 0.18.0 - Transformers: 4.52.3 - Pytorch: 2.6.0 - Datasets: 3.6.0 - Tokenizers: 0.21.1
Qwen2.5-14B-Open-R1-GRPO
Llama-2-70b-chat-longlora-32k
Qwen2.5-3B-Open-R1-GRPO
This model is a fine-tuned version of Qwen/Qwen2.5-3B-Instruct on the open-r1/OpenR1-Math-220k dataset. It has been trained using TRL. This model was trained with GRPO, a method introduced in DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models. - TRL: 0.18.0 - Transformers: 4.52.3 - Pytorch: 2.6.0 - Datasets: 3.6.0 - Tokenizers: 0.21.1
Llama-2-70b-longlora-32k
Llama-2-13b-longlora-32k
LongAlpaca-7B-16k
Llama-2-7b-longlora-8k-ft
Llama-2-13b-longlora-16k
Llama-2-13b-longlora-18k-ft
zephyr-7b-sft-full
Qwen2.5-32B-Open-R1-GRPO
This model is a fine-tuned version of None. It has been trained using TRL. This model was trained with GRPO, a method introduced in DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models. - TRL: 0.21.0 - Transformers: 4.52.3 - Pytorch: 2.7.0 - Datasets: 3.6.0 - Tokenizers: 0.21.4