lightblue

57 models • 5 total models in database
Sort by:

suzume-llama-3-8B-multilingual

NaNK
llama
14,245
113

suzume-llama-3-8B-multilingual-orpo-borda-half

NaNK
llama
9,110
16

suzume-llama-3-8B-multilingual-orpo-borda-top25

NaNK
llama
8,634
3

suzume-llama-3-8B-multilingual-orpo-borda-top75

NaNK
llama
7,926
4

suzume-llama-3-8B-multilingual-orpo-borda-full

NaNK
llama
7,896
2

japanese-mpt-7b

NaNK
license:apache-2.0
1,343
23

suzume-llama-3-8B-japanese-gguf

NaNK
base_model:meta-llama/Meta-Llama-3-8B-Instruct
184
12

suzume-llama-3-8B-multilingual-gguf

NaNK
base_model:meta-llama/Meta-Llama-3-8B-Instruct
178
27

suzume-llama-3-8B-japanese

NaNK
llama
118
24

lb-reranker-0.5B-v1.0

NaNK
license:apache-2.0
113
74

DeepSeek-R1-Distill-Qwen-7B-Multilingual

NaNK
dataset:lightblue/reasoning-multilingual-R1-Llama-70B-train
66
22

DeepSeek-R1-Distill-Qwen-14B-Multilingual

NaNK
dataset:lightblue/reasoning-multilingual-R1-Llama-70B-train
47
13

suzume-llama-3-8B-multilingual-orpo-borda-half-gguf

NaNK
license:cc-by-nc-4.0
36
10

Karasu-DPO-7B

NaNK
license:apache-2.0
27
4

DeepSeek-R1-Distill-Qwen-7B-Japanese

NaNK
llama-factory
25
31

DeepSeek-R1-Distill-Qwen-1.5B-Multilingual

NaNK
dataset:lightblue/reasoning-multilingual-R1-Llama-70B-train
25
24

karasu-1.1B

NaNK
llama
15
7

Karasu-Mixtral-8x22B-v0.1-gguf

NaNK
10
14

qarasu-14B-chat-plus-unleashed

NaNK
9
21

karasu-7B-chat

NaNK
license:apache-2.0
6
2

kurage-es

license:apache-2.0
6
1

reranker_0.5_bin_filt

NaNK
llama-factory
5
1

Karasu-Mixtral-8x22B-v0.1

This is a finetune of the newly released mistral-community/Mixtral-8x22B-v0.1 base model. As the base model has not explicitly been trained to chat, we trained this model on a multilingual chat dataset so that the LLM community can use this model for conversations. The accuracy of the model is surprisingly high, and has a decently fast inference speed (roughly 40 tokens/s single batch on our tests), so we believe this will be useful to the community. We have tested (and thus recommend) running this model on vLLM. We recommend running it from the vLLM openAI server, using the following command: which is how we ran it on a 4 x A100 (80GB) machine. You can then call this model from Python installing the openai package: We will be uploading a 4bit AWQ model soon to make it easier to run this model on other machines (watch this space!). From qualitative testing, the model seems pretty smart, especially in English, and has very good recall of facts. It can still get confused with some logical questions, but has also passed a lot of the logical questions I have thrown at it that other open source LLMs often fail. FAIL! The banana would still be in the kitchen, as I put the plate on the banana. Sort of a failure, I'd prefer to say "the center of the Earth". The idea is a bit original. but the casting is 2/3rds Dune actresses. We trained this model on conversations between human users and GPT-4. 6,206 conversations from the openchat/openchatsharegpt4dataset dataset (link) 3,011 conversations that we created. We wanted to increase the representation of nonenglish prompts in our training dataset, so we sampled initial prompts from lmsys/lmsys-chat-1m, stratifying based on language. We then prompted gpt-4-0125 with these, and used the results as training data. We plan to release more information on this second dataset soon, as we are using it another dataset. The complete data used to train this model can be found at lightblue/gpt4conversationsmultilingual We trained this model using Axolotl's 4bit QLoRA configuration for roughly 100 minutes in a A100 (80GB) x 4 environment on the Azure cloud (StandardNC96adsA100v4). We used Deepspeed Zero2 to effectively train over 4 GPUs.

NaNK
license:apache-2.0
4
62

karasu-7B-chat-plus

NaNK
license:apache-2.0
4
2

kurage-sw

license:apache-2.0
4
1

karasu-7B

NaNK
license:apache-2.0
3
3

kurage-en

license:apache-2.0
3
3

kurage-ru

license:apache-2.0
3
3

kurage-vi

license:apache-2.0
3
1

qwen2.5-7B-instruct-orpo

NaNK
llama-factory
3
1

qwen2.5-7B-instruct-simpo

NaNK
llama-factory
3
0

reranker_0.5_cont_filt

NaNK
llama-factory
3
0

kurage-multilingual

license:apache-2.0
2
28

openorca_stx

llama
2
16

karasu-7B-chat-plus-unleashed

NaNK
license:apache-2.0
2
8

kurage-ja

license:apache-2.0
2
6

kurage-ko

license:apache-2.0
2
4

jod

license:apache-2.0
2
1

kurage-ar

license:apache-2.0
2
1

kurage-hi

license:apache-2.0
2
1

Karasu-Mixtral-8x22B-v0.1-gguf-test

NaNK
2
0

reranker_0.5_cont

NaNK
llama-factory
2
0

Jamba-v0.1-chat-multilingual

NaNK
license:apache-2.0
1
23

ao-karasu-72B

NaNK
1
14

ao-karasu-72B-AWQ-4bit

NaNK
1
2

kurage-zh

license:apache-2.0
1
2

tinyllama_chat_jsquad

NaNK
base_model:TinyLlama/TinyLlama-1.1B-Chat-v0.3
1
0

qwen2.5-7B-instruct-kto

NaNK
llama-factory
1
0

qwen2.5-7B-instruct-orpo2

NaNK
llama-factory
1
0

reranker_0.5_bincont_filt

NaNK
llama-factory
1
0

reranker_0.5_cont_filt_7max

NaNK
llama-factory
1
0

Karasu-Mixtral-8x22B-v0.1-AWQ

NaNK
0
7

shitsu_text_scorer

license:mit
0
7

kurage-id

license:apache-2.0
0
2

all_search_calling_4bit

NaNK
0
1

kurage-th

license:apache-2.0
0
1

lb-reranker-0.5B-v1.0-rev

NaNK
license:apache-2.0
0
1