arcee-ai

175 models • 3 total models in database
Sort by:

Trinity-Large-Preview-W4A16

license:apache-2.0
17,080
6

Trinity-Mini-GGUF

license:apache-2.0
14,553
42

AFM-4.5B-Base

AFM-4.5B-Base is a 4.5 billion parameter instruction-tuned model developed by Arcee.ai, designed for enterprise-grade performance across diverse deployment environments from cloud to edge. The base model was trained on a dataset of 8 trillion tokens, comprising 6.5 trillion tokens of general pretraining data followed by 1.5 trillion tokens of midtraining data with enhanced focus on mathematical reasoning and code generation. Following pretraining, the model underwent supervised fine-tuning on high-quality instruction datasets. The instruction-tuned model was further refined through reinforcement learning on verifiable rewards as well as for human preference. We use a modified version of TorchTitan for pretraining, Axolotl for supervised fine-tuning, and a modified version of Verifiers for reinforcement learning. The development of AFM-4.5B prioritized data quality as a fundamental requirement for achieving robust model performance. We collaborated with DatologyAI, a company specializing in large-scale data curation. DatologyAI's curation pipeline integrates a suite of proprietary algorithms—model-based quality filtering, embedding-based curation, target distribution-matching, source mixing, and synthetic data. Their expertise enabled the creation of a curated dataset tailored to support strong real-world performance. The model architecture follows a standard transformer decoder-only design based on Vaswani et al., incorporating several key modifications for enhanced performance and efficiency. Notable architectural features include grouped query attention for improved inference efficiency and ReLU^2 activation functions instead of SwiGLU to enable sparsification while maintaining or exceeding performance benchmarks. The model available in this repo is the base model following merging and context extension. Model Architecture: ArceeForCausalLM Parameters: 4.5B Training Tokens: 8T License: Apache-2.0 You can use the model directly with the `transformers` library.

NaNK
license:apache-2.0
10,364
31

Trinity-Large-Thinking

license:apache-2.0
8,191
137

Trinity-Mini

license:apache-2.0
2,630
139

AFM-4.5B

AFM-4.5B is a 4.5 billion parameter instruction-tuned model developed by Arcee.ai, designed for enterprise-grade performance across diverse deployment environments from cloud to edge. The base model was trained on a dataset of 8 trillion tokens, comprising 6.5 trillion tokens of general pretraining data followed by 1.5 trillion tokens of midtraining data with enhanced focus on mathematical reasoning and code generation. Following pretraining, the model underwent supervised fine-tuning on high-quality instruction datasets. The instruction-tuned model was further refined through reinforcement learning on verifiable rewards as well as for human preference. We use a modified version of TorchTitan for pretraining, Axolotl for supervised fine-tuning, and a modified version of Verifiers for reinforcement learning. The development of AFM-4.5B prioritized data quality as a fundamental requirement for achieving robust model performance. We collaborated with DatologyAI, a company specializing in large-scale data curation. DatologyAI's curation pipeline integrates a suite of proprietary algorithms—model-based quality filtering, embedding-based curation, target distribution-matching, source mixing, and synthetic data. Their expertise enabled the creation of a curated dataset tailored to support strong real-world performance. The model architecture follows a standard transformer decoder-only design based on Vaswani et al., incorporating several key modifications for enhanced performance and efficiency. Notable architectural features include grouped query attention for improved inference efficiency and ReLU^2 activation functions instead of SwiGLU to enable sparsification while maintaining or exceeding performance benchmarks. The model available in this repo is the instruct model following supervised fine-tuning and reinforcement learning. View our documentation here for more details: https://docs.arcee.ai/arcee-foundation-models/introduction-to-arcee-foundation-models Model Architecture: ArceeForCausalLM Parameters: 4.5B Training Tokens: 8T License: Apache 2.0 Recommended settings: temperature: 0.5 topk: 50 topp: 0.95 repeatpenalty: 1.1 Qwen3 and SmolLM's reasoning approach causes their scores to vary wildly from suite to suite - but these are all scores on our internal harness with the same hyperparameters. Be sure to reference their reported scores. SmolLM just released its bench. You can use the model directly with the `transformers` library. We recommend a lower temperature, around 0.5, for optimal performance. You can access this model directly via the Together Playground. Support for llama.cpp and Intel OpenVINO is available:

NaNK
license:apache-2.0
1,327
86

Trinity-Nano-Base

license:apache-2.0
1,307
20

SuperNova-Medius-GGUF

license:apache-2.0
1,050
63

Trinity-Large-Preview

license:apache-2.0
995
168

Virtuoso-Medium-v2-GGUF

NaNK
license:apache-2.0
721
11

Llama-3.1-SuperNova-Lite-GGUF

NaNK
base_model:meta-llama/Llama-3.1-8B-Instruct
607
17

Arcee-VyLinh

NaNK
license:apache-2.0
594
41

Arcee-Nova-GGUF

462
14

Homunculus-GGUF

license:apache-2.0
418
14

AFM-4.5B-GGUF

AFM-4.5B is a 4.5 billion parameter instruction-tuned model developed by Arcee.ai, designed for enterprise-grade performance across diverse deployment environments from cloud to edge. The base model was trained on a dataset of 8 trillion tokens, comprising 6.5 trillion tokens of general pretraining data followed by 1.5 trillion tokens of midtraining data with enhanced focus on mathematical reasoning and code generation. Following pretraining, the model underwent supervised fine-tuning on high-quality instruction datasets. The instruction-tuned model was further refined through reinforcement learning on verifiable rewards as well as for human preference. We use a modified version of TorchTitan for pretraining, Axolotl for supervised fine-tuning, and a modified version of Verifiers for reinforcement learning. The development of AFM-4.5B prioritized data quality as a fundamental requirement for achieving robust model performance. We collaborated with DatologyAI, a company specializing in large-scale data curation. DatologyAI's curation pipeline integrates a suite of proprietary algorithms—model-based quality filtering, embedding-based curation, target distribution-matching, source mixing, and synthetic data. Their expertise enabled the creation of a curated dataset tailored to support strong real-world performance. The model architecture follows a standard transformer decoder-only design based on Vaswani et al., incorporating several key modifications for enhanced performance and efficiency. Notable architectural features include grouped query attention for improved inference efficiency and ReLU^2 activation functions instead of SwiGLU to enable sparsification while maintaining or exceeding performance benchmarks. The model available in this repo is the instruct model following supervised fine-tuning and reinforcement learning. Model Architecture: ArceeForCausalLM Parameters: 4.5B Training Tokens: 8T License: Apache 2.0 Recommended settings: temperature: 0.5 topk: 50 topp: 0.95 repeatpenalty: 1.1 Qwen3 and SmolLM's reasoning approach causes their scores to vary wildly from suite to suite - but these are all scores on our internal harness with the same hyperparameters. Be sure to reference their reported scores. SmolLM just released its bench.

NaNK
license:apache-2.0
383
28

Virtuoso-Small-GGUF

NaNK
license:apache-2.0
335
9

Llama-Spark-GGUF

license:llama3.1
328
8

Arcee-SuperNova-v1

Arcee-SuperNova-v1 (70B) is a merged model built from multiple advanced training approaches. At its core is a distilled version of Llama-3.1-405B-Instruct into Llama-3.1-70B-Instruct, using out DistillKit to preserve instruction-following strengths while reducing size. Alongside this, another Llama-3.1-70B model was instruction-tuned using synthetic data from our Evol-Kit pipeline, improving precision and adherence across diverse queries. Updates were integrated mid-epoch for smoother performance gains. A third version underwent Direct Preference Optimization (DPO) to better align with human feedback. While its contribution was smaller, it helped refine final alignment. The resulting Arcee-SuperNova combines all three, delivering strong human preference alignment and state-of-the-art instruction-following ability. - Architecture Base: Llama-3.1-70B-Instruct - Parameter Count: 70B - License: [Llama3] - General intelligence and instruction following - Serving as a base to be retrained over time using Reinforcement Learning from Human Feedback (RLHF) - Mathematical applications and queries Arcee-SuperNova-v1 (70B) is released under the Llama-3 license. You are free to use, modify, and distribute this model in both commercial and non-commercial applications, subject to the terms and conditions of the license. If you have questions or would like to share your experiences using Arcee-SuperNova-v1 (70B), please connect with us on social media. We’re excited to see what you build—and how this model helps you innovate!

NaNK
llama
317
16

Virtuoso-Large-GGUF

Virtuoso-Large (72B) is our most powerful and versatile general-purpose model, designed to excel at handling complex and varied tasks across domains. With state-of-the-art performance, it offers unparalleled capability for nuanced understanding, contextual adaptability, and high accuracy. - Architecture Base: Qwen2.5-72B - Parameter Count: 72B - License: qwen - Advanced content creation, such as technical writing and creative storytelling - Data summarization and report generation for cross-functional domains - Detailed knowledge synthesis and deep-dive insights from diverse datasets - Multilingual support for international operations and communications Virtuoso-Large (72B) is released under the qwen License. If you have questions or would like to share your experiences using Virtuoso-Large (72B), please connect with us on social media. We’re excited to see what you build—and how this model helps you innovate!

310
6

Llama-3-SEC-Chat

llama
270
37

arcee-lite-GGUF

license:apache-2.0
259
9

Arcee-Blitz-GGUF

license:apache-2.0
258
9

Arcee-Spark-GGUF

license:apache-2.0
240
27

Arcee-SuperNova-v1-GGUF

NaNK
license:llama3
218
2

Trinity-Mini-Base

license:apache-2.0
191
18

KidRails

llama
179
1

Trinity-Large-Preview-FP8-Block

license:apache-2.0
162
1

Caller-GGUF

license:apache-2.0
158
4

Arcee-Maestro-7B-Preview-GGUF

NaNK
license:apache-2.0
158
3

Llama-3-SEC-Base

llama
146
12

Mistral-7B-Instruct-v0.2-sliced-24-layer

NaNK
144
7

Arcee-Scribe

license:apache-2.0
142
15

Arcee-Scribe-GGUF

license:apache-2.0
142
11

DeepSeek-V3-0324-bf16

license:mit
140
3

DeepSeek-R1-bf16

138
15

Biomistral-Calme-Instruct-7b

NaNK
license:apache-2.0
138
1

Meraj-Mini

license:apache-2.0
129
17

Trinity-Large-TrueBase

license:apache-2.0
126
62

Arcee-Agent

license:apache-2.0
113
93

Llama-3-SEC-Chat-GGUF

license:llama3
113
9

Trinity-Nano-Preview-W4A16

license:apache-2.0
109
1

gemma-3b-it-expanded

NaNK
license:apache-2.0
94
1

arcee-lite

license:apache-2.0
91
24

Virtuoso-Medium-v2

NaNK
license:apache-2.0
85
56

Virtuoso-Large

NaNK
84
29

gemma-7b-slerp

NaNK
license:apache-2.0
84
1

gemma-7b-zephyr-alpaca-it-ties

NaNK
license:apache-2.0
83
2

Hermes-Mistral-Legal-Slerp

NaNK
license:apache-2.0
83
0

Qwen2.5-32B-Instruct-FP8

NaNK
83
0

saul-mistral-v0.2-7b-slerp

NaNK
license:apache-2.0
82
0

GLM-4-32B-Base-32K

NaNK
license:mit
81
37

Virtuoso-Lite-GGUF

NaNK
81
7

Patent-Instruct-7b

NaNK
llama
81
1

Biomistral-Exp-Slerp

NaNK
license:apache-2.0
81
1

CS-Calme-Instruct-7b

NaNK
license:apache-2.0
81
0

saul-zephyr-7b-slerp

NaNK
license:apache-2.0
81
0

Saul-Nous-Hermes-2-Mistral-7B-DPO-slerp

NaNK
license:apache-2.0
81
0

Llama-3-OpenBioLLM-JSL-8B-SLERP

NaNK
llama
81
0

Mistral-Instruct-Orca-Slerp

NaNK
license:apache-2.0
80
1

SEC-MBX-7B-DPO

NaNK
license:apache-2.0
80
1

arcee-blitz-caller-beta

NaNK
license:apache-2.0
80
1

Mistral-Hermes-Support-Ties

NaNK
license:apache-2.0
80
0

Calme-Instruct-Extended

NaNK
license:apache-2.0
80
0

Patent-Instruct-Pro

NaNK
llama
80
0

Patent-Instruct-LLaMA-Pro

NaNK
llama
80
0

saul-zephyr-7b-ties

NaNK
license:apache-2.0
80
0

Saul-Nous-Hermes-2-Mistral-7B-DPO-Ties

NaNK
license:apache-2.0
79
2

sec-mistral-7b-instruct-v2

NaNK
79
1

Gemma-merged-2B-ties

NaNK
license:apache-2.0
79
0

Saul-Base-Calme-7B-Instruct-slerp

NaNK
license:apache-2.0
79
0

PMC_LLaMA_Vicuna_13B_Slerp

NaNK
llama
79
0

Mistral-7B-Instruct-v0.2-expanded

NaNK
78
4

Mistral-7B-Instruct-v0.2-expanded-sec-1.6B-tokens

NaNK
78
4

Legal-Saul-Multiverse-7b

NaNK
license:apache-2.0
78
2

saul-mistral-instruct-v0.1-7b-ties

NaNK
license:apache-2.0
78
1

SEC-1.6-Calme-7B-Instruct

NaNK
license:apache-2.0
78
1

Llama-3-Base-Instruct-Slerp

NaNK
llama
78
1

Mistral-Lora-Adapter-CS-Slerp

NaNK
license:apache-2.0
78
0

gemma-7b-it-zaphyr-slerp

NaNK
license:apache-2.0
78
0

Gemma-Openchat-SauerkrautLM

NaNK
license:apache-2.0
78
0

Customer-Support-Clown-7b

NaNK
license:apache-2.0
78
0

Clown-Saul-Extended

license:apache-2.0
78
0

Patent-Instruct-Extended-40

NaNK
llama
78
0

arcee-sec-mistral-7b

NaNK
78
0

Saul-Instruct-Mistral-7B-Instruct-v0.2-Slerp

NaNK
license:apache-2.0
78
0

Llama-3-MegaMed-8B-Model-Stock

NaNK
llama
78
0

Hermes-2-Pro-WizardMath-7B-SLERP

NaNK
78
0

Caller

Caller (32B) is a robust model engineered for seamless integrations and optimized for managing complex tool-based interactions and API function calls. Its strength lies in precise execution, intelligent orchestration, and effective communication between systems, making it indispensable for sophisticated automation pipelines. - Architecture Base: Qwen2.5-32B - Parameter Count: 32B - License: Apache-2.0 - Managing integrations between CRMs, ERPs, and other enterprise systems - Running multi-step workflows with intelligent condition handling - Orchestrating external tool interactions like calendar scheduling, email parsing, or data extraction - Real-time monitoring and diagnostics in IoT or SaaS environments GGUF format available here License Caller (32B) is released under the Apache-2.0 License. You are free to use, modify, and distribute this model in both commercial and non-commercial applications, subject to the terms and conditions of the license. If you have questions or would like to share your experiences using Caller (32B), please connect with us on social media. We’re excited to see what you build—and how this model helps you innovate!

NaNK
license:apache-2.0
77
12

Trinity-Tokenizer

77
3

Virtuoso-Lite-PreDistill

NaNK
llama
77
3

Hermes-Mistral-Saul-Slerp

NaNK
license:apache-2.0
77
1

Saul-Legal-Calme-Instruct

NaNK
license:apache-2.0
77
1

gemma-10b-it-expanded

NaNK
license:apache-2.0
77
1

llama_from_mistral_instruct_v2

llama
77
1

saul-mistral-v0.1-7b-slerp

NaNK
license:apache-2.0
77
1

Saul-Base-Clown-7B-Instruct-slerp

NaNK
license:apache-2.0
77
1

Patent-Llama-7B-Chat-Slerp

NaNK
llama
77
1

Patent-Base-7b

NaNK
llama
77
0

gemma-7b-alpaca-zaphyr-slerp

NaNK
license:apache-2.0
77
0

Saul-Instruct-Clown-7b

NaNK
license:apache-2.0
77
0

Calme-Clown-Extended

license:apache-2.0
77
0

Patent-Instruct-Extended

NaNK
llama
77
0

Patent-Instruct-Barcenas-Orca-2

NaNK
llama
77
0

Patent-Base-InternLM2-7B-Ties

NaNK
llama
77
0

mistral-v2-sec-dolphin

NaNK
77
0

Patent-Base-Llama-2-Chat-7B-Slerp

NaNK
llama
77
0

MyAlee-Qwen-Instruct-v2-16k-v1

NaNK
license:apache-2.0
77
0

Arcee-Spark-FP32

license:apache-2.0
77
0

BioMistral-merged-zephyr

NaNK
76
1

Biomistral-Clown-Slerp

NaNK
license:apache-2.0
76
1

SEC-1.6-MBX-7B-DPO

NaNK
license:apache-2.0
76
1

Saul-Instruct-Extended

license:apache-2.0
76
0

Patent-Instruct-Orca-2-Model-Stock

llama
76
0

Patent-Instruct-Internlm2-7B-Ties

NaNK
llama
76
0

Patent-Base-Orca-2-7B-Slerp

NaNK
llama
76
0

Llama-3-8B-Instruct-Base-Slerp

NaNK
llama
76
0

Clown-DPO-Extended

NaNK
license:apache-2.0
75
5

MedLLaMA-Vicuna-13B-Slerp

NaNK
llama
75
3

sec-mistral-v2-Hercules

NaNK
75
1

myalee-v3-L31-8B

NaNK
llama
75
1

MistralProSupportSlerp

NaNK
license:apache-2.0
75
0

Alpaca-Dragon-Smaug-Slerp

NaNK
llama
75
0

mistral-sliced

license:apache-2.0
75
0

Customer-Support-Clown-Extended

NaNK
license:apache-2.0
75
0

zilo-instruct-v2-sft-filtered

NaNK
license:apache-2.0
75
0

BioMistral-merged-instruct

NaNK
74
2

Gemma-Zephyr-Dolly-Chat-Slerp

NaNK
license:apache-2.0
74
0

Patent-Base-Orca-2-7B-Ties

NaNK
llama
74
0

Patent-Instruct-Llama-2-Chat-7B-Slerp

NaNK
llama
74
0

Llama-3-Medical-JSL-WiNGPT2-SLERP

NaNK
llama
74
0

Homunculus

NaNK
license:apache-2.0
73
98

WitchLM-1.5B

NaNK
license:apache-2.0
70
7

SEC-Calme-7B-Instruct

NaNK
license:apache-2.0
70
1

teeny-tiny-mixtral

70
1

sec-mistral-7b-instruct-1.6-epoch

NaNK
70
0

patent-evol-merge

llama
70
0

cpt-16B-auto-sft-ties-post-merge-auto-dpo

NaNK
llama
70
0

gemma-7b-alpaca-it-ties

NaNK
license:apache-2.0
69
1

Patent-Instruct-Orca-2

NaNK
llama
69
1

AFM-4.5B-Preview

NaNK
license:apache-2.0
68
5

Patent-Base-Barcenas-Orca-2-7B-Slerp

NaNK
llama
68
0

Arcee-Blitz-AWQ

license:apache-2.0
65
6

SuperNova-Medius-FP8

62
2

Meraj-Mini-FP8

62
1

deepseek-v2-chat-0628-awq

60
6

Llama-3.1-SuperNova-Lite-FP8

llama
59
1

Trinity-Large-Base

license:apache-2.0
56
53

AFM 4.5B Base Pre Anneal

AFM-4.5B-Base-Pre-Anneal is a 4.5 billion parameter instruction-tuned model developed by Arcee.ai, designed for enterprise-grade performance across diverse deployment environments from cloud to edge. The base model was trained on a dataset of 6.5 trillion tokens of general pretraining data. We use a modified version of TorchTitan for pretraining. The development of AFM-4.5B prioritized data quality as a fundamental requirement for achieving robust model performance. We collaborated with DatologyAI, a company specializing in large-scale data curation. DatologyAI's curation pipeline integrates a suite of proprietary algorithms—model-based quality filtering, embedding-based curation, target distribution-matching, source mixing, and synthetic data. Their expertise enabled the creation of a curated dataset tailored to support strong real-world performance. The model architecture follows a standard transformer decoder-only design based on Vaswani et al., incorporating several key modifications for enhanced performance and efficiency. Notable architectural features include grouped query attention for improved inference efficiency and ReLU^2 activation functions instead of SwiGLU to enable sparsification while maintaining or exceeding performance benchmarks. The model available in this repo is the base model before it was annealed with math and code and before merging and context extension. Model Architecture: ArceeForCausalLM Parameters: 4.5B Training Tokens: 6.5T - this model is pre-annealing with math and code and uses only the general dataset. License: Apache-2.0 You can use the model directly with the `transformers` library.

NaNK
license:apache-2.0
53
3

Arcee-VyLinh-GGUF

45
3

Trinity-Mini-W4A16

license:apache-2.0
42
2

llama-8b-sft-qlora

NaNK
llama
34
0

Virtuoso-Lite-4bit-mlx

The Model mlx-community/Virtuoso-Lite-4bit was converted to MLX format from arcee-ai/Virtuoso-Lite using mlx-lm version 0.21.1.

NaNK
llama
33
0

Virtuoso-Medium-v2-3bit-mlx

NaNK
license:apache-2.0
33
0

Virtuoso-Lite-3bit-mlx

The Model mlx-community/Virtuoso-Lite-3bit was converted to MLX format from arcee-ai/Virtuoso-Lite using mlx-lm version 0.21.1.

NaNK
llama
32
0

Virtuoso-Lite-6bit-mlx

NaNK
llama
32
0

Virtuoso-Medium-v2-6bit-mlx

NaNK
license:apache-2.0
32
0

Virtuoso-Medium-v2-bf16-mlx

NaNK
license:apache-2.0
31
0

Arcee-Nova-AWQ

30
2

Virtuoso-Lite-8bit-mlx

NaNK
llama
30
0

Virtuoso-Medium-v2-4bit-mlx

NaNK
license:apache-2.0
30
0

Virtuoso-Medium-v2-8bit-mlx

NaNK
license:apache-2.0
30
0

WitchLM-1.5B-GGUF

NaNK
29
2

Arcee-Maestro-7B-Preview-AWQ

NaNK
license:apache-2.0
29
1

zilo-sft-qlora

NaNK
license:apache-2.0
18
0

Trinity-Nano-Base-Pre-Anneal

license:apache-2.0
16
8

Trinity-Mini-Base-Pre-Anneal

license:apache-2.0
13
9

Trinity-Large-Thinking-NVFP4

license:apache-2.0
13
0

Trinity-Large-Preview-FP8

license:apache-2.0
0
13

AFM-4.5B-ov

This model repository contains 16-bit, 8-bit, and 4-bit versions of AFM-4.5B optimized with the Intel OpenVINO toolkit. The original model was converted with Intel OpenVINO 2025.03 and Hugging Face Optimum Intel. You can easily try the model with this code snippet: AFM-4.5B is a 4.5 billion parameter instruction-tuned model developed by Arcee.ai, designed for enterprise-grade performance across diverse deployment environments from cloud to edge. The base model was trained on a dataset of 8 trillion tokens, comprising 6.5 trillion tokens of general pretraining data followed by 1.5 trillion tokens of midtraining data with enhanced focus on mathematical reasoning and code generation. Following pretraining, the model underwent supervised fine-tuning on high-quality instruction datasets. The instruction-tuned model was further refined through reinforcement learning on verifiable rewards as well as for human preference. We use a modified version of TorchTitan for pretraining, Axolotl for supervised fine-tuning, and a modified version of Verifiers for reinforcement learning. The development of AFM-4.5B prioritized data quality as a fundamental requirement for achieving robust model performance. We collaborated with DatologyAI, a company specializing in large-scale data curation. DatologyAI's curation pipeline integrates a suite of proprietary algorithms—model-based quality filtering, embedding-based curation, target distribution-matching, source mixing, and synthetic data. Their expertise enabled the creation of a curated dataset tailored to support strong real-world performance. The model architecture follows a standard transformer decoder-only design based on Vaswani et al., incorporating several key modifications for enhanced performance and efficiency. Notable architectural features include grouped query attention for improved inference efficiency and ReLU^2 activation functions instead of SwiGLU to enable sparsification while maintaining or exceeding performance benchmarks. The model available in this repo is the instruct model following supervised fine-tuning and reinforcement learning. Model Architecture: ArceeForCausalLM Parameters: 4.5B Training Tokens: 8T License: Apache 2.0 Recommended settings: temperature: 0.5 topk: 50 topp: 0.95 repeatpenalty: 1.1 Qwen3 and SmolLM's reasoning approach causes their scores to vary wildly from suite to suite - but these are all scores on our internal harness with the same hyperparameters. Be sure to reference their reported scores. SmolLM just released its bench.

NaNK
license:apache-2.0
0
7

AFM-4.5B-Base-KDA-Only

NaNK
license:apache-2.0
0
5

AFM-4.5B-Base-KDA-NoPE

NaNK
license:apache-2.0
0
5

Trinity-Nano-Preview

license:apache-2.0
0
3

Trinity-Large-Thinking-GGUF

license:apache-2.0
0
1

Trinity-Nano-Preview-NVFP4

license:apache-2.0
0
1

Trinity-Large-Preview-GGUF

license:apache-2.0
0
1