Sakalti
Saka-14B
Base model: sometimesanotion Lamarck 14B v0.7, sometimesanotion Qwenvergence 14B v11.
ultiima-72B
Language support for English and Chinese.
iturkaAI-xsmall
Qwen2.5-1B-Instruct
Base model: Sakalti SJT 0.5B library name: transformers
model-4-novem
nekomata-7b-gguf
light-7b-beta
Base model: Qwen 2.5 7B Instruct.
Saka-7.6B
Base model: suayptalha/HomerCreativeAnvita-Mix-Qw7B, SousiOmine/Kuroiso-CR-7B-20250124.
ultiima-32B
Apache 2.0 licensed library using transformers.
ultiima-125B
Built With Qwen Details このモデルは80レイヤーから140レイヤーに増やしたモデルです。性能は低下しているのでファインチューニングしてよりいい精度を目指す用途を想定しています。 merge This is a merge of pre-trained language models created using mergekit. This model was merged using the passthrough merge method. The following models were included in the merge: Sakalti/ultiima-72B The following YAML configuration was used to produce this model:
Saka-3.8B
Saka-1.5B
Base model: SakanaAI TinySwallow 1.5B Instruct, Sakalti SJT 1.5B Alpha.
deneb-v1-7b
Magro-7b-v1.1
This is a merge of pre-trained language models created using mergekit. This model was merged using the TIES merge method using HuggingFaceH4/zephyr-7b-alpha as a base. The following models were included in the merge: Sakalti/magro-7B The following YAML configuration was used to produce this model:
LunarPass-1
Saba1.5-1.5B
License: Apache 2.0, Library Name: Transformers, Inference: True
model-3
Library name: transformers, tags: mergekit.
model-4-novem2
llama-3-yanyuedao-8b-Instruct
Saba1.5-Pro
トレーニングには15m 23sかかりました。 ライセンスはapache2.0です。 Uploaded model - Developed by: Sakalti - License: apache2.0 - Finetuned from model : Sakalti/Saba1.5-1.5B This qwen model was trained 2x faster with Unsloth and Huggingface's TRL library.
ultiima-14B
Base model: Qwen Qwen2.5 14B Qwen Qwen2.5 14B Instruct.
ultiima-78B
This is a merge of pre-trained language models created using mergekit. This model was merged using the TIES merge method using Saxo/Linkbricks-Horizon-AI-Avengers-V2-78B as a base. The following models were included in the merge: MaziyarPanahi/calme-3.2-instruct-78b The following YAML configuration was used to produce this model:
SakaMoe-3x14B-Instruct
Model Details This model is a reincarnation of Saka-14B as a Mixture of Experts (MoE) and has approximately 45 billion parameters. Model Description - Developed by: [Sakalti] - Funded by [optional]: [More Information Needed] - Shared by [optional]: [More Information Needed] - Model type: [MoE] - Language(s) (NLP): [Japanese, English..] - License: [apache-2.0 license] - Finetuned from model [optional]: [More Information Needed] - Repository: [More Information Needed] - Paper [optional]: [More Information Needed] - Demo [optional]: [More Information Needed] Bias, Risks, and Limitations 1.This model supports Japanese and English, so unexpected responses may occur in other languages. Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: [More Information Needed] - Hours used: [More Information Needed] - Cloud Provider: [More Information Needed] - Compute Region: [More Information Needed] - Carbon Emitted: [More Information Needed]
Saba1.5-Pro-3B
Base model: Saba1.5-Pro library name: transformers.
light-3B-beta
This is a merge of pre-trained language models created using mergekit. This model was merged using the TIES merge method using Qwen/Qwen2.5-3B as a base. The following models were included in the merge: Qwen/Qwen2.5-3B-Instruct The following YAML configuration was used to produce this model:
SJT-8B
Base model Sakalti/model-3 library name transformers.
flatum
Lunar-4B
This project utilizes HODACHI/Borea-Phi-3.5-mini-Instruct-Common, a model based on Phi-3.5-mini-Instruct and fine-tuned by Axcxept co., ltd. merge This is a merge of pre-trained language models created using mergekit. This model was merged using the TIES merge method using microsoft/phi-3.5-mini-instruct as a base. The following models were included in the merge: AXCXEPT/Borea-Phi-3.5-mini-Instruct-Common The following YAML configuration was used to produce this model:
Saka-1B
EleutherAI-pythia-14m-ONNX
mont-xl-agi
SJT-1.7B
Text generation inference model based on Sakalti/Qwen2.5-test-2.
SJT-7.5B
Base model from the PrithivMLmods Taurus Opus 7B library using transformers.
Ultiima-78B-v2
Saba2-14B-Preview
Base model includes arcee-ai Virtuoso Small and arcee-ai SuperNova Medius.
Euphrates-14B
Base model includes Spestly Athena 1 14B and Qwen Qwen 2.5 14B.
Anemoi-3B
Base model: Qwen Qwen2.5 3B Instruct, bunnycore Qwen 2.5 3B Rp lora model.
fira
SJT-1.5B-Alpha
Base model Qwen Qwen2.5 1.5B Instruct Qwen Qwen2.5 1.5B.
sakalinear-1-1.0
oxyge1-33B
Saba1-1.8B
Base model: micaebe/Qwen2.5-1.5B-Instruct-QwQ, Sakalti/model-4.
magro-7B
Base model: HuggingfaceH4/zephyr-7b-beta. Inference: true. Tags:
Neptuno-3B
Base model: Spestly Athena 1 3B, Sakalti Light 3B.
Neptuno-Alpha
Base model includes bunnycore Qwen 2.5 3B RP Mix and Sakalti Neptuno 3B.
SJT-1.5B-Alpha-1.1
Base model includes Sakalti Saba 1.5 Pro and Sakalti SJT 1.5B Alpha.
CodeSaka-1.1-3.8B
hotalai
gpt-neox-small
model-1
Kan1-2.5b
Saba1.5-1.7B
QwenTest-7
Base model Qwen 2.5 0.5B Instruct library name transformers.
SJTPass-5
Base model: Sakalti/SJT-0.5B, library name: transformers.
Saka-7.2B
Core purpose is text generation inference. Base model is llm-jp/llm-jp-3-7.2b-instruct.
Template-4
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: [More Information Needed] - Funded by [optional]: [More Information Needed] - Shared by [optional]: [More Information Needed] - Model type: [More Information Needed] - Language(s) (NLP): [More Information Needed] - License: [More Information Needed] - Finetuned from model [optional]: [More Information Needed] - Repository: [More Information Needed] - Paper [optional]: [More Information Needed] - Demo [optional]: [More Information Needed] Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: [More Information Needed] - Hours used: [More Information Needed] - Cloud Provider: [More Information Needed] - Compute Region: [More Information Needed] - Carbon Emitted: [More Information Needed]
SJT-RP-Lora-1.5B
Template-5
model-4
Saba1-1.8B-Coder
light-1.1-3B
Base model: Sakalti/light-3B inference: true tags:
SJT-14B
Base model: djuna Q2.5 Veltha 14B, hotmailuser QwenSlerp2 14B.
Qwen2.5-test-2
SakalFusion-7B-Alpha
Base model: AIDC-AI/marco-o1, Qwen/Qwen2.5-7B.
Lunar-Bfloat16-4B
ultiima-14B-v0.3
Base model includes sometimesanotion/Qwenvergence-14B-v9 and sometimesanotion/Qwen2.5-14B-Vimarckoso-v3.
SJT-24B-Alpha
Base model: Sakalti/ultiima-14B-v0.4 library name: transformers.
SJT-7B-V1.1
Base model includes Suayptalha HomerCreativeAnvita Mix Qw7B and Bunnycore Qwen 2.5 7B RRP 1M.
SJT-2B-V1.1
Base model: rinna/gemma-2-Baku-2b-it, prithivMLmods/GWQ2b.
Saka-3.8B-Coder
SakaMoe-3x1.6B-Instruct
This model is licensed under Apache 2.0 and supports the English language.
qwen2.5
iturkaAI-large
Qwen7b
mont-normal
sansan
This is a merge of pre-trained language models created using mergekit. This model was merged using the Model Stock merge method using Qwen/Qwen2.5-7B-Instruct as a base. The following models were included in the merge: Isaak-Carter/Josiefied-Qwen2.5-7B-Instruct-abliterated-v2 ClaudioItaly/Qwen2.5-Boster FourOhFour/Vapor7B huihui-ai/Qwen2.5-7B-Instruct-abliterated Qwen/Qwen2.5-Math-7B The following YAML configuration was used to produce this model:
model-6
Saba-Passthrough-2
Base model: Saba1.5-Pro, library name: transformers.
Phi3.5-Comets-3.8B
Base model: win10/Phi-3.5-mini-instruct-24-9-29, FreedomIntelligence/Apollo2-3.8B.
Saba2-3B
Base model for text generation inference.
Llama3.2-3B-Uranus-1
Text generation inference model based on AXCXEPT/EZO-Llama-3.2-3B-Instruct.
tara-3.8B
Core purpose is text generation. Base model is unsloth/Phi-3.5-mini-instruct.
SJT-4B
Text generation inference model based on Sakalti/Tara-3.8B-v1.1.
SJT-0.5B
Base model Qwen Qwen2.5 0.5B Instruct.
SJT-2B
Base model: google/gemma-2-2b-jpn-it, library name: transformers.
qwen2.5-2.3B
Text generation model using the Transformers library with an Apache 2.0 license.
SJT-3.7B
Base model: llm-jp/llm-jp-3-3.7b-instruct, llm-jp/llm-jp-3-3.7b-Instruct.
ultiima-108B
SJT-2.4B
Base model: Sakalti/SJT-1.5B-Alpha library name: transformers.
SJTPass-2
Base model: Sakalti/SJT-0.5B library name: transformers.
ultiima-72B-v1.5
Base model: Sakalti/ultiima-72B, shuttleai/shuttle-3.
SakalFusion-7B-Beta
Base model includes Qwen 2.5 Coder 7B Instruct and Qwen 2.5 Math 7B Instruct.
SJT-990M
ultiima-78B-Q2-mlx
SJT-8B-V1.1
Base model: Sakalti/SJT-7B-V1.1 library name: transformers.
SakaSlerp-7.6B
Mystica1-4B
- Developed by: Sakalti - License: apache-2.0 - Finetuned from model : unsloth/Qwen3-4B This qwen3 model was trained 2x faster with Unsloth and Huggingface's TRL library.
light-3B
Text generation inference using transformers.
ultiima-14B-v0.2
Base model: sometimesanotion/Lamarck-14B-v0.7, sometimesanotion/Lamarck-14B-v0.6.
SakaSlerp-14B
beril
lakeland
Saba1-7B
Base model AIDC-AI Marco-o1 Sakalti model 3.
Saba1-3B
This is a merge of pre-trained language models created using mergekit. This model was merged using the passthrough merge method. The following models were included in the merge: Qwen/Qwen2.5-1.5B-Instruct The following YAML configuration was used to produce this model:
Saba2-1.7B
Qwen2.5-1.5B-Instruct-1.25
This modelcard aims to be a base template for new models. It has been generated using this raw template. Model Details - Developed by: [Sakalti] - Funded by [optional]: [More Information Needed] - Shared by [optional]: [More Information Needed] - Model type: [text generation] - Language(s) (NLP): [english etc..] - License: [Apache license 2.0] - Finetuned from model [optional]: [More Information Needed] - Repository: [More Information Needed] - Paper [optional]: [More Information Needed] - Demo [optional]: [More Information Needed] Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: [More Information Needed] - Hours used: [More Information Needed] - Cloud Provider: [More Information Needed] - Compute Region: [More Information Needed] - Carbon Emitted: [More Information Needed]
Kani-3B
Tara-3.8B-v1.1
Text generation inference model based on Sakalti/Tara-3.8B.
Qwen2.5-7B-Mini
bonsai-2b
SJT-4B-v1.1
SJT-2.4B-Alpha
- Developed by: Sakalti - License: apache-2.0 - Finetuned from model : Sakalti/SJT-2.4B This qwen2 model was trained 2x faster with Unsloth and Huggingface's TRL library.
mine-3B
mine-1.5B
SJT-7B-1M
SJT-Moe2x7.5B
This model is based on Sakalti/SJT-7.5B and is licensed under the Apache 2.0 license.
ultiima-14B-v0.4
Base model: sometimesanotion/Lamarck-14B-v0.7, sometimesanotion/LoRA-256-Base-Qwenvergence.