lmsys

40 models • 3 total models in database
Sort by:

vicuna-7b-v1.5

--- inference: false license: llama2 ---

NaNK
llama
194,632
374

sglang-EAGLE-llama2-chat-7B

This model is copied from https://huggingface.co/yuhuili/EAGLE-llama2-chat-7B with the following modifications to make it compatible with SGLang: - Modify the architecture in config.json to LlamaFo...

NaNK
llama
61,237
2

gpt-oss-20b-bf16

gpt-oss-20b-bf16 Model Introduction This model is the bf16 version converted from openai/gpt-oss-20b. Usage You can use this model in SGLang with the following instructions. Installation For more details https://github.com/sgl-project/sglang/issues/8833

NaNK
52,646
10

gpt-oss-120b-bf16

gpt-oss-120b-bf16 Model Introduction This model is the bf16 version converted from openai/gpt-oss-120b. Usage You can use this model in SGLang with the following instructions. Installation For more details https://github.com/sgl-project/sglang/issues/8833

NaNK
30,990
3

vicuna-13b-v1.5-16k

NaNK
llama
24,704
225

vicuna-13b-v1.5

Vicuna is a chat assistant trained by fine-tuning Llama 2 on user-shared conversations collected from ShareGPT. - Developed by: LMSYS - Model type: An auto-regressive language model based on the transformer architecture - License: Llama 2 Community License Agreement - Finetuned from model: Llama 2 - Repository: https://github.com/lm-sys/FastChat - Blog: https://lmsys.org/blog/2023-03-30-vicuna/ - Paper: https://arxiv.org/abs/2306.05685 - Demo: https://chat.lmsys.org/ The primary use of Vicuna is research on large language models and chatbots. The primary intended users of the model are researchers and hobbyists in natural language processing, machine learning, and artificial intelligence. - Command line interface: https://github.com/lm-sys/FastChat#vicuna-weights - APIs (OpenAI API, Huggingface API): https://github.com/lm-sys/FastChat/tree/main#api Vicuna v1.5 is fine-tuned from Llama 2 with supervised instruction fine-tuning. The training data is around 125K conversations collected from ShareGPT.com. See more details in the "Training Details of Vicuna Models" section in the appendix of this paper. Vicuna is evaluated with standard benchmarks, human preference, and LLM-as-a-judge. See more details in this paper and leaderboard.

NaNK
llama
23,925
235

sglang-EAGLE3-LLaMA3.1-Instruct-8B

NaNK
llama
16,380
2

longchat-13b-16k

NaNK
llama
12,373
134

vicuna-7b-v1.3

NOTE: New version available Please check out a newer version of the weights here. Vicuna is a chat assistant trained by fine-tuning LLaMA on user-shared conversations collected from ShareGPT. - Developed by: LMSYS - Model type: An auto-regressive language model based on the transformer architecture. - License: Non-commercial license - Finetuned from model: LLaMA. - Repository: https://github.com/lm-sys/FastChat - Blog: https://lmsys.org/blog/2023-03-30-vicuna/ - Paper: https://arxiv.org/abs/2306.05685 - Demo: https://chat.lmsys.org/ The primary use of Vicuna is research on large language models and chatbots. The primary intended users of the model are researchers and hobbyists in natural language processing, machine learning, and artificial intelligence. - Command line interface: https://github.com/lm-sys/FastChat#vicuna-weights. - APIs (OpenAI API, Huggingface API): https://github.com/lm-sys/FastChat/tree/main#api. Vicuna v1.3 is fine-tuned from LLaMA with supervised instruction fine-tuning. The training data is around 125K conversations collected from ShareGPT.com. See more details in the "Training Details of Vicuna Models" section in the appendix of this paper. Vicuna is evaluated with standard benchmarks, human preference, and LLM-as-a-judge. See more details in this paper and leaderboard. Difference between different versions of Vicuna See vicunaweightsversion.md

NaNK
llama
10,418
139

sglang-EAGLE-LLaMA3-Instruct-8B

This model is copied from https://huggingface.co/yuhuili/EAGLE-LLaMA3-Instruct-8B with the following modifications to make it compatible with SGLang: - Modify the architecture in config.json to LlamaForCausalLMEagle

NaNK
llama
8,188
2

vicuna-13b-v1.3

NOTE: New version available Please check out a newer version of the weights here. Vicuna is a chat assistant trained by fine-tuning LLaMA on user-shared conversations collected from ShareGPT. - Developed by: LMSYS - Model type: An auto-regressive language model based on the transformer architecture. - License: Non-commercial license - Finetuned from model: LLaMA. - Repository: https://github.com/lm-sys/FastChat - Blog: https://lmsys.org/blog/2023-03-30-vicuna/ - Paper: https://arxiv.org/abs/2306.05685 - Demo: https://chat.lmsys.org/ The primary use of Vicuna is research on large language models and chatbots. The primary intended users of the model are researchers and hobbyists in natural language processing, machine learning, and artificial intelligence. - Command line interface: https://github.com/lm-sys/FastChat#vicuna-weights. - APIs (OpenAI API, Huggingface API): https://github.com/lm-sys/FastChat/tree/main#api. Vicuna v1.3 is fine-tuned from LLaMA with supervised instruction fine-tuning. The training data is around 125K conversations collected from ShareGPT.com. See more details in the "Training Details of Vicuna Models" section in the appendix of this paper. Vicuna is evaluated with standard benchmarks, human preference, and LLM-as-a-judge. See more details in this paper and leaderboard. Difference between different versions of Vicuna See vicunaweightsversion.md

NaNK
llama
3,622
201

longchat-7b-v1.5-32k

NaNK
llama
2,894
60

vicuna-7b-v1.1

NaNK
llama
2,889
77

EAGLE3-gpt-oss-120b-bf16

NaNK
llama
2,386
6

vicuna-33b-v1.3

Vicuna is a chat assistant trained by fine-tuning LLaMA on user-shared conversations collected from ShareGPT. - Developed by: LMSYS - Model type: An auto-regressive language model based on the transformer architecture. - License: Non-commercial license - Finetuned from model: LLaMA. - Repository: https://github.com/lm-sys/FastChat - Blog: https://lmsys.org/blog/2023-03-30-vicuna/ - Paper: https://arxiv.org/abs/2306.05685 - Demo: https://chat.lmsys.org/ The primary use of Vicuna is research on large language models and chatbots. The primary intended users of the model are researchers and hobbyists in natural language processing, machine learning, and artificial intelligence. - Command line interface: https://github.com/lm-sys/FastChat#vicuna-weights. - APIs (OpenAI API, Huggingface API): https://github.com/lm-sys/FastChat/tree/main#api. Vicuna v1.3 is fine-tuned from LLaMA with supervised instruction fine-tuning. The training data is around 125K conversations collected from ShareGPT.com. See more details in the "Training Details of Vicuna Models" section in the appendix of this paper. Vicuna is evaluated with standard benchmarks, human preference, and LLM-as-a-judge. See more details in this paper and leaderboard. Difference between different versions of Vicuna See vicunaweightsversion.md

NaNK
llama
1,896
293

vicuna-7b-v1.5-16k

NaNK
llama
1,711
84

vicuna-13b-delta-v1.1

NOTE: New version available Please check out a newer version of the weights here. NOTE: This "delta model" cannot be used directly. Users have to apply it on top of the original LLaMA weights to get actual Vicuna weights. See instructions. Vicuna is a chat assistant trained by fine-tuning LLaMA on user-shared conversations collected from ShareGPT. - Developed by: LMSYS - Model type: An auto-regressive language model based on the transformer architecture. - License: Non-commercial license - Finetuned from model: LLaMA. - Repository: https://github.com/lm-sys/FastChat - Blog: https://lmsys.org/blog/2023-03-30-vicuna/ - Paper: https://arxiv.org/abs/2306.05685 - Demo: https://chat.lmsys.org/ The primary use of Vicuna is research on large language models and chatbots. The primary intended users of the model are researchers and hobbyists in natural language processing, machine learning, and artificial intelligence. Command line interface: https://github.com/lm-sys/FastChat#vicuna-weights. APIs (OpenAI API, Huggingface API): https://github.com/lm-sys/FastChat/tree/main#api. Vicuna v1.1 is fine-tuned from LLaMA with supervised instruction fine-tuning. The training data is around 70K conversations collected from ShareGPT.com. See more details in the "Training Details of Vicuna Models" section in the appendix of this paper. Vicuna is evaluated with standard benchmarks, human preference, and LLM-as-a-judge. See more details in this paper and leaderboard. Difference between different versions of Vicuna See vicunaweightsversion.md

NaNK
llama
930
410

vicuna-7b-delta-v1.1

NaNK
llama
928
202

vicuna-13b-v1.1

NaNK
llama
921
99

DeepSeek-V3-NextN

870
2

DeepSeek-R1-NextN

705
6

sglang-EAGLE3-Llama-4-Scout-17B-16E-Instruct-v1

NaNK
llama
698
0

Qwen3-235B-A22B-EAGLE3

Model Introduction The Eagle3 draft model was trained using the SpecForge framework for the Qwen3-235B-A22B model, leveraging a combination of UltraChat and ShareGPT datasets. Usage You can use this Eagle3 draft model in SGLang with the following command:

NaNK
llama
660
4

longchat-7b-16k

NaNK
llama
532
49

sglang-EAGLE3-LLaMA3.3-Instruct-70B

NaNK
llama
442
1

fastchat-t5-3b-v1.0

NaNK
license:apache-2.0
325
369

toxicchat-t5-large-v1.0

license:apache-2.0
259
6

vicuna-7b-delta-v0

NaNK
llama
242
165

vicuna-13b-delta-v0

NOTE: New version available Please check out a newer version of the weights here. NOTE: This "delta model" cannot be used directly. Users have to apply it on top of the original LLaMA weights to get actual Vicuna weights. See instructions. Vicuna is a chat assistant trained by fine-tuning LLaMA on user-shared conversations collected from ShareGPT. - Developed by: LMSYS - Model type: An auto-regressive language model based on the transformer architecture. - License: Non-commercial license - Finetuned from model: LLaMA. - Repository: https://github.com/lm-sys/FastChat - Blog: https://lmsys.org/blog/2023-03-30-vicuna/ - Paper: https://arxiv.org/abs/2306.05685 - Demo: https://chat.lmsys.org/ The primary use of Vicuna is research on large language models and chatbots. The primary intended users of the model are researchers and hobbyists in natural language processing, machine learning, and artificial intelligence. Command line interface: https://github.com/lm-sys/FastChat#vicuna-weights. APIs (OpenAI API, Huggingface API): https://github.com/lm-sys/FastChat/tree/main#api. Vicuna v0 is fine-tuned from LLaMA with supervised instruction fine-tuning. The training data is around 70K conversations collected from ShareGPT.com. See more details in the "Training Details of Vicuna Models" section in the appendix of this paper. Vicuna is evaluated with standard benchmarks, human preference, and LLM-as-a-judge. See more details in this paper and leaderboard. Difference between different versions of Vicuna See vicunaweightsversion.md

NaNK
llama
180
453

DeepSeek-V3-0324-NextN

140
0

SGLang-EAGLE3-Qwen3-30B-A3B-Instruct-2507-SpecForge-Nex

NaNK
llama
52
0

SGLang-EAGLE3-Llama-3.3-70B-Instruct-SpecForge

NaNK
llama
49
0

SGLang-EAGLE3-Qwen3-Coder-480B-A35B-Instruct-SpecForge-EigenAI

NaNK
llama
43
0

sglang-EAGLE-LLaMA3-Instruct-70B

This model is copied from https://huggingface.co/yuhuili/EAGLE-LLaMA3-Instruct-70B with the following modifications to make it compatible with SGLang: - Modify the architecture in config.json to LlamaForCausalLMEagle

NaNK
llama
40
0

sglang-EAGLE3-Llama-4-Maverick-17B-128E-Instruct-v1

NaNK
llama
36
0

sglang-EAGLE-llama2-chat-70B

This model is copied from https://huggingface.co/yuhuili/EAGLE-llama2-chat-70B with the following modifications to make it compatible with SGLang: - Modify the architecture in config.json to LlamaForCausalLMEagle

NaNK
llama
30
1

SGLang-EAGLE3-Qwen3-235B-A22B-Instruct-2507-SpecForge-Meituan

NaNK
llama
24
0

SGLang-EAGLE3-Llama-4-Scout-17B-16E-Instruct-SpecForge

NaNK
llama
22
0

SGLang-EAGLE3-Llama-3.1-8B-Instruct-SpecForge

NaNK
llama
22
0

SGLang-EAGLE3-Qwen3-Coder-30B-A3B-Instruct-SpecForge

NaNK
llama
9
0