Sakalti

120 models • 60 total models in database
Sort by:

Saka-14B

Base model: sometimesanotion Lamarck 14B v0.7, sometimesanotion Qwenvergence 14B v11.

NaNK
1,342
8

ultiima-72B

Language support for English and Chinese.

NaNK
145
3

iturkaAI-xsmall

105
0

Qwen2.5-1B-Instruct

Base model: Sakalti SJT 0.5B library name: transformers

NaNK
license:apache-2.0
59
0

model-4-novem

NaNK
43
0

nekomata-7b-gguf

NaNK
llama-cpp-python
21
0

light-7b-beta

Base model: Qwen 2.5 7B Instruct.

NaNK
license:apache-2.0
13
1

Saka-7.6B

Base model: suayptalha/HomerCreativeAnvita-Mix-Qw7B, SousiOmine/Kuroiso-CR-7B-20250124.

NaNK
12
1

ultiima-32B

Apache 2.0 licensed library using transformers.

NaNK
license:apache-2.0
9
6

ultiima-125B

Built With Qwen Details このモデルは80レイヤーから140レイヤーに増やしたモデルです。性能は低下しているのでファインチューニングしてよりいい精度を目指す用途を想定しています。 merge This is a merge of pre-trained language models created using mergekit. This model was merged using the passthrough merge method. The following models were included in the merge: Sakalti/ultiima-72B The following YAML configuration was used to produce this model:

NaNK
9
1

Saka-3.8B

NaNK
8
1

Saka-1.5B

Base model: SakanaAI TinySwallow 1.5B Instruct, Sakalti SJT 1.5B Alpha.

NaNK
8
1

deneb-v1-7b

NaNK
8
1

Magro-7b-v1.1

This is a merge of pre-trained language models created using mergekit. This model was merged using the TIES merge method using HuggingFaceH4/zephyr-7b-alpha as a base. The following models were included in the merge: Sakalti/magro-7B The following YAML configuration was used to produce this model:

NaNK
license:mit
8
0

LunarPass-1

NaNK
7
1

Saba1.5-1.5B

License: Apache 2.0, Library Name: Transformers, Inference: True

NaNK
license:apache-2.0
7
0

model-3

Library name: transformers, tags: mergekit.

NaNK
7
0

model-4-novem2

NaNK
7
0

llama-3-yanyuedao-8b-Instruct

NaNK
llama
7
0

Saba1.5-Pro

トレーニングには15m 23sかかりました。 ライセンスはapache2.0です。 Uploaded model - Developed by: Sakalti - License: apache2.0 - Finetuned from model : Sakalti/Saba1.5-1.5B This qwen model was trained 2x faster with Unsloth and Huggingface's TRL library.

NaNK
license:apache-2.0
6
3

ultiima-14B

Base model: Qwen Qwen2.5 14B Qwen Qwen2.5 14B Instruct.

NaNK
license:apache-2.0
6
1

ultiima-78B

This is a merge of pre-trained language models created using mergekit. This model was merged using the TIES merge method using Saxo/Linkbricks-Horizon-AI-Avengers-V2-78B as a base. The following models were included in the merge: MaziyarPanahi/calme-3.2-instruct-78b The following YAML configuration was used to produce this model:

NaNK
6
0

SakaMoe-3x14B-Instruct

Model Details This model is a reincarnation of Saka-14B as a Mixture of Experts (MoE) and has approximately 45 billion parameters. Model Description - Developed by: [Sakalti] - Funded by [optional]: [More Information Needed] - Shared by [optional]: [More Information Needed] - Model type: [MoE] - Language(s) (NLP): [Japanese, English..] - License: [apache-2.0 license] - Finetuned from model [optional]: [More Information Needed] - Repository: [More Information Needed] - Paper [optional]: [More Information Needed] - Demo [optional]: [More Information Needed] Bias, Risks, and Limitations 1.This model supports Japanese and English, so unexpected responses may occur in other languages. Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: [More Information Needed] - Hours used: [More Information Needed] - Cloud Provider: [More Information Needed] - Compute Region: [More Information Needed] - Carbon Emitted: [More Information Needed]

NaNK
license:apache-2.0
6
0

Saba1.5-Pro-3B

Base model: Saba1.5-Pro library name: transformers.

NaNK
license:apache-2.0
5
1

light-3B-beta

This is a merge of pre-trained language models created using mergekit. This model was merged using the TIES merge method using Qwen/Qwen2.5-3B as a base. The following models were included in the merge: Qwen/Qwen2.5-3B-Instruct The following YAML configuration was used to produce this model:

NaNK
5
1

SJT-8B

Base model Sakalti/model-3 library name transformers.

NaNK
5
1

flatum

NaNK
5
0

Lunar-4B

This project utilizes HODACHI/Borea-Phi-3.5-mini-Instruct-Common, a model based on Phi-3.5-mini-Instruct and fine-tuned by Axcxept co., ltd. merge This is a merge of pre-trained language models created using mergekit. This model was merged using the TIES merge method using microsoft/phi-3.5-mini-instruct as a base. The following models were included in the merge: AXCXEPT/Borea-Phi-3.5-mini-Instruct-Common The following YAML configuration was used to produce this model:

NaNK
license:mit
4
1

Saka-1B

NaNK
license:apache-2.0
4
1

EleutherAI-pythia-14m-ONNX

license:apache-2.0
4
0

mont-xl-agi

NaNK
4
0

SJT-1.7B

Text generation inference model based on Sakalti/Qwen2.5-test-2.

NaNK
license:apache-2.0
4
0

SJT-7.5B

Base model from the PrithivMLmods Taurus Opus 7B library using transformers.

NaNK
llama
4
0

Ultiima-78B-v2

NaNK
4
0

Saba2-14B-Preview

Base model includes arcee-ai Virtuoso Small and arcee-ai SuperNova Medius.

NaNK
license:apache-2.0
3
2

Euphrates-14B

Base model includes Spestly Athena 1 14B and Qwen Qwen 2.5 14B.

NaNK
3
2

Anemoi-3B

Base model: Qwen Qwen2.5 3B Instruct, bunnycore Qwen 2.5 3B Rp lora model.

NaNK
3
1

fira

license:apache-2.0
3
0

SJT-1.5B-Alpha

Base model Qwen Qwen2.5 1.5B Instruct Qwen Qwen2.5 1.5B.

NaNK
license:apache-2.0
3
0

sakalinear-1-1.0

NaNK
3
0

oxyge1-33B

NaNK
license:apache-2.0
2
2

Saba1-1.8B

Base model: micaebe/Qwen2.5-1.5B-Instruct-QwQ, Sakalti/model-4.

NaNK
license:apache-2.0
2
1

magro-7B

Base model: HuggingfaceH4/zephyr-7b-beta. Inference: true. Tags:

NaNK
license:mit
2
1

Neptuno-3B

Base model: Spestly Athena 1 3B, Sakalti Light 3B.

NaNK
2
1

Neptuno-Alpha

Base model includes bunnycore Qwen 2.5 3B RP Mix and Sakalti Neptuno 3B.

NaNK
2
1

SJT-1.5B-Alpha-1.1

Base model includes Sakalti Saba 1.5 Pro and Sakalti SJT 1.5B Alpha.

NaNK
license:apache-2.0
2
1

CodeSaka-1.1-3.8B

NaNK
2
1

hotalai

license:cc-by-sa-4.0
2
0

gpt-neox-small

license:mit
2
0

model-1

NaNK
2
0

Kan1-2.5b

NaNK
2
0

Saba1.5-1.7B

NaNK
license:apache-2.0
2
0

QwenTest-7

Base model Qwen 2.5 0.5B Instruct library name transformers.

NaNK
license:apache-2.0
2
0

SJTPass-5

Base model: Sakalti/SJT-0.5B, library name: transformers.

NaNK
2
0

Saka-7.2B

Core purpose is text generation inference. Base model is llm-jp/llm-jp-3-7.2b-instruct.

NaNK
llama
2
0

Template-4

This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: [More Information Needed] - Funded by [optional]: [More Information Needed] - Shared by [optional]: [More Information Needed] - Model type: [More Information Needed] - Language(s) (NLP): [More Information Needed] - License: [More Information Needed] - Finetuned from model [optional]: [More Information Needed] - Repository: [More Information Needed] - Paper [optional]: [More Information Needed] - Demo [optional]: [More Information Needed] Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: [More Information Needed] - Hours used: [More Information Needed] - Cloud Provider: [More Information Needed] - Compute Region: [More Information Needed] - Carbon Emitted: [More Information Needed]

2
0

SJT-RP-Lora-1.5B

NaNK
2
0

Template-5

2
0

model-4

NaNK
license:apache-2.0
1
1

Saba1-1.8B-Coder

NaNK
1
1

light-1.1-3B

Base model: Sakalti/light-3B inference: true tags:

NaNK
1
1

SJT-14B

Base model: djuna Q2.5 Veltha 14B, hotmailuser QwenSlerp2 14B.

NaNK
license:apache-2.0
1
1

Qwen2.5-test-2

NaNK
license:apache-2.0
1
1

SakalFusion-7B-Alpha

Base model: AIDC-AI/marco-o1, Qwen/Qwen2.5-7B.

NaNK
1
1

Lunar-Bfloat16-4B

NaNK
license:mit
1
1

ultiima-14B-v0.3

Base model includes sometimesanotion/Qwenvergence-14B-v9 and sometimesanotion/Qwen2.5-14B-Vimarckoso-v3.

NaNK
1
1

SJT-24B-Alpha

Base model: Sakalti/ultiima-14B-v0.4 library name: transformers.

NaNK
1
1

SJT-7B-V1.1

Base model includes Suayptalha HomerCreativeAnvita Mix Qw7B and Bunnycore Qwen 2.5 7B RRP 1M.

NaNK
1
1

SJT-2B-V1.1

Base model: rinna/gemma-2-Baku-2b-it, prithivMLmods/GWQ2b.

NaNK
1
1

Saka-3.8B-Coder

NaNK
1
1

SakaMoe-3x1.6B-Instruct

This model is licensed under Apache 2.0 and supports the English language.

NaNK
license:apache-2.0
1
0

qwen2.5

NaNK
1
0

iturkaAI-large

license:apache-2.0
1
0

Qwen7b

NaNK
license:apache-2.0
1
0

mont-normal

NaNK
1
0

sansan

This is a merge of pre-trained language models created using mergekit. This model was merged using the Model Stock merge method using Qwen/Qwen2.5-7B-Instruct as a base. The following models were included in the merge: Isaak-Carter/Josiefied-Qwen2.5-7B-Instruct-abliterated-v2 ClaudioItaly/Qwen2.5-Boster FourOhFour/Vapor7B huihui-ai/Qwen2.5-7B-Instruct-abliterated Qwen/Qwen2.5-Math-7B The following YAML configuration was used to produce this model:

NaNK
1
0

model-6

NaNK
1
0

Saba-Passthrough-2

Base model: Saba1.5-Pro, library name: transformers.

license:apache-2.0
1
0

Phi3.5-Comets-3.8B

Base model: win10/Phi-3.5-mini-instruct-24-9-29, FreedomIntelligence/Apollo2-3.8B.

NaNK
license:mit
1
0

Saba2-3B

Base model for text generation inference.

NaNK
1
0

Llama3.2-3B-Uranus-1

Text generation inference model based on AXCXEPT/EZO-Llama-3.2-3B-Instruct.

NaNK
llama
1
0

tara-3.8B

Core purpose is text generation. Base model is unsloth/Phi-3.5-mini-instruct.

NaNK
llama
1
0

SJT-4B

Text generation inference model based on Sakalti/Tara-3.8B-v1.1.

NaNK
llama
1
0

SJT-0.5B

Base model Qwen Qwen2.5 0.5B Instruct.

NaNK
license:apache-2.0
1
0

SJT-2B

Base model: google/gemma-2-2b-jpn-it, library name: transformers.

NaNK
1
0

qwen2.5-2.3B

Text generation model using the Transformers library with an Apache 2.0 license.

NaNK
license:apache-2.0
1
0

SJT-3.7B

Base model: llm-jp/llm-jp-3-3.7b-instruct, llm-jp/llm-jp-3-3.7b-Instruct.

NaNK
llama
1
0

ultiima-108B

NaNK
1
0

SJT-2.4B

Base model: Sakalti/SJT-1.5B-Alpha library name: transformers.

NaNK
1
0

SJTPass-2

Base model: Sakalti/SJT-0.5B library name: transformers.

NaNK
license:apache-2.0
1
0

ultiima-72B-v1.5

Base model: Sakalti/ultiima-72B, shuttleai/shuttle-3.

NaNK
1
0

SakalFusion-7B-Beta

Base model includes Qwen 2.5 Coder 7B Instruct and Qwen 2.5 Math 7B Instruct.

NaNK
1
0

SJT-990M

NaNK
license:apache-2.0
1
0

ultiima-78B-Q2-mlx

NaNK
1
0

SJT-8B-V1.1

Base model: Sakalti/SJT-7B-V1.1 library name: transformers.

NaNK
1
0

SakaSlerp-7.6B

NaNK
1
0

Mystica1-4B

- Developed by: Sakalti - License: apache-2.0 - Finetuned from model : unsloth/Qwen3-4B This qwen3 model was trained 2x faster with Unsloth and Huggingface's TRL library.

NaNK
license:apache-2.0
1
0

light-3B

Text generation inference using transformers.

NaNK
0
2

ultiima-14B-v0.2

Base model: sometimesanotion/Lamarck-14B-v0.7, sometimesanotion/Lamarck-14B-v0.6.

NaNK
0
2

SakaSlerp-14B

NaNK
0
2

beril

NaNK
0
1

lakeland

NaNK
0
1

Saba1-7B

Base model AIDC-AI Marco-o1 Sakalti model 3.

NaNK
0
1

Saba1-3B

This is a merge of pre-trained language models created using mergekit. This model was merged using the passthrough merge method. The following models were included in the merge: Qwen/Qwen2.5-1.5B-Instruct The following YAML configuration was used to produce this model:

NaNK
0
1

Saba2-1.7B

NaNK
license:apache-2.0
0
1

Qwen2.5-1.5B-Instruct-1.25

This modelcard aims to be a base template for new models. It has been generated using this raw template. Model Details - Developed by: [Sakalti] - Funded by [optional]: [More Information Needed] - Shared by [optional]: [More Information Needed] - Model type: [text generation] - Language(s) (NLP): [english etc..] - License: [Apache license 2.0] - Finetuned from model [optional]: [More Information Needed] - Repository: [More Information Needed] - Paper [optional]: [More Information Needed] - Demo [optional]: [More Information Needed] Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: [More Information Needed] - Hours used: [More Information Needed] - Cloud Provider: [More Information Needed] - Compute Region: [More Information Needed] - Carbon Emitted: [More Information Needed]

NaNK
license:apache-2.0
0
1

Kani-3B

NaNK
0
1

Tara-3.8B-v1.1

Text generation inference model based on Sakalti/Tara-3.8B.

NaNK
llama
0
1

Qwen2.5-7B-Mini

NaNK
license:apache-2.0
0
1

bonsai-2b

NaNK
0
1

SJT-4B-v1.1

NaNK
license:mit
0
1

SJT-2.4B-Alpha

- Developed by: Sakalti - License: apache-2.0 - Finetuned from model : Sakalti/SJT-2.4B This qwen2 model was trained 2x faster with Unsloth and Huggingface's TRL library.

NaNK
license:apache-2.0
0
1

mine-3B

NaNK
0
1

mine-1.5B

NaNK
license:apache-2.0
0
1

SJT-7B-1M

NaNK
0
1

SJT-Moe2x7.5B

This model is based on Sakalti/SJT-7.5B and is licensed under the Apache 2.0 license.

NaNK
license:apache-2.0
0
1

ultiima-14B-v0.4

Base model: sometimesanotion/Lamarck-14B-v0.7, sometimesanotion/LoRA-256-Base-Qwenvergence.

NaNK
0
1

SakaDare-3.8B

NaNK
0
1

SakaRP-1-1.5B

NaNK
license:apache-2.0
0
1

SakaPass-1

NaNK
0
1