janhq

105 models • 1 total models in database
Sort by:

Jan-v1-4B-GGUF

NaNK
license:apache-2.0
82,338
131

Jan-v1-2509-gguf

[](https://github.com/menloresearch/deep-research) [](https://opensource.org/licenses/Apache-2.0) [](https://jan.ai/) We have released a small weight update, jan-v1-2509, which refines the original v1. No architectural changes. Slightly lower performance on SimpleQA compared to jan-v1. Slightly mproved results on other chat benchmarks and overall more reliable Jan-v1 is the first release in the Jan Family, designed for agentic reasoning and problem-solving within the Jan App. Based on our Lucy model, Jan-v1 achieves improved performance through model scaling. Jan-v1 uses the Qwen3-4B-thinking model to provide enhanced reasoning capabilities and tool utilization. This architecture delivers better performance on complex agentic tasks. Question Answering (SimpleQA) For question-answering, Jan-v1 shows a significant performance gain from model scaling, achieving 91.1% accuracy. The 91.1% SimpleQA accuracy with Jan-v1 remains a highlight, though Jan-v1-2509 focuses on balancing factual QA with improved reliability across chat-based reasoning tasks. These benchmarks evaluate the model's conversational and instructional capabilities. Jan-v1 is optimized for direct integration with the Jan App. Simply select the model from the Jan App interface for immediate access to its full capabilities. - Discussions: HuggingFace Community - Jan App: Learn more about the Jan App at jan.ai () Note By default we have system prompt in chat template, this is to make sure the model having the same performance with the benchmark result. You can also use the vanilla chat template without system prompt in the file chattemplateraw.jinja.

license:apache-2.0
31,647
12

Jan-v2-VL-high-gguf

license:apache-2.0
22,880
23

Jan-v1-edge-gguf

Jan-v1-edge: Distilled for Edge, Built for Web Search [](https://github.com/menloresearch/deep-research) [](https://opensource.org/licenses/Apache-2.0) [](https://jan.ai/) Jan-v1-edge is a lightweight agentic model built for fast, reliable on-device execution. As the second release in the Jan Family, it is distilled from the larger `Jan-v1` model, preserving strong reasoning and problem-solving ability in a smaller footprint suitable for resource-constrained environments. Jan-v1-edge was developed through a two-phase post-training process. The first phase, Supervised Fine-Tuning (SFT), transferred core capabilities from the `Jan-v1` teacher model to the smaller student. The second phase, Reinforcement Learning with Verifiable Rewards (RLVR) —the same method used in `Jan-v1` and `Lucy`—further optimized reasoning efficiency, tool use, and correctness. This staged approach delivers reliable results on complex, interactive workloads. Despite having only 1.7B parameters, Jan-v1-edge achieves 83% accuracy—nearly matching the larger Jan-nano-128k—demonstrating its efficiency and robustness. Versus Qwen 3 1.7B Thinking, Jan-v1-edge shows a slight degradation on instruction-following and CreativeWriting, while remaining comparable or better on EQBench and recency QA. Jan-v1-edge is optimized for direct integration with the Jan App. Simply select the model from the Jan App interface for immediate access to its full capabilities. - Discussions: HuggingFace Community - Jan App: Discover more about the Jan App at jan.ai

license:apache-2.0
7,919
18

Jan-v2-VL-low

license:apache-2.0
5,454
15

Jan-code-4b-gguf

NaNK
license:apache-2.0
2,241
19

Jan-v1-4B

NaNK
license:apache-2.0
1,419
344

Jan-v3-4B-base-instruct

NaNK
license:apache-2.0
927
40

Jan-v1-2509

[](https://github.com/menloresearch/deep-research) [](https://opensource.org/licenses/Apache-2.0) [](https://jan.ai/) We have released a small weight update, jan-v1-2509, which refines the original v1. No architectural changes. Slightly lower performance on SimpleQA compared to jan-v1. Slightly mproved results on other chat benchmarks and overall more reliable Jan-v1 is the first release in the Jan Family, designed for agentic reasoning and problem-solving within the Jan App. Based on our Lucy model, Jan-v1 achieves improved performance through model scaling. Jan-v1 uses the Qwen3-4B-thinking model to provide enhanced reasoning capabilities and tool utilization. This architecture delivers better performance on complex agentic tasks. Question Answering (SimpleQA) For question-answering, Jan-v1 shows a significant performance gain from model scaling, achieving 91.1% accuracy. The 91.1% SimpleQA accuracy represents a significant milestone in factual question answering for models of this scale, demonstrating the effectiveness of our scaling and fine-tuning approach. These benchmarks evaluate the model's conversational and instructional capabilities. Jan-v1 is optimized for direct integration with the Jan App. Simply select the model from the Jan App interface for immediate access to its full capabilities. - Discussions: HuggingFace Community - Jan App: Learn more about the Jan App at jan.ai () Note By default we have system prompt in chat template, this is to make sure the model having the same performance with the benchmark result. You can also use the vanilla chat template without system prompt in the file chattemplateraw.jinja.

NaNK
license:apache-2.0
522
32

Vistral-7b-Chat-GGUF

NaNK
222
10

Jan-v2-VL-med-gguf

license:apache-2.0
179
4

Jan-v1-edge

Jan-v1-edge: Distilled for Edge, Built for Web Search [](https://github.com/menloresearch/deep-research) [](https://opensource.org/licenses/Apache-2.0) [](https://jan.ai/) Jan-v1-edge is a lightweight agentic model built for fast, reliable on-device execution. As the second release in the Jan Family, it is distilled from the larger `Jan-v1` model, preserving strong reasoning and problem-solving ability in a smaller footprint suitable for resource-constrained environments. Jan-v1-edge was developed through a two-phase post-training process. The first phase, Supervised Fine-Tuning (SFT), transferred core capabilities from the `Jan-v1` teacher model to the smaller student. The second phase, Reinforcement Learning with Verifiable Rewards (RLVR) —the same method used in `Jan-v1` and `Lucy`—further optimized reasoning efficiency, tool use, and correctness. This staged approach delivers reliable results on complex, interactive workloads. Despite having only 1.7B parameters, Jan-v1-edge achieves 83% accuracy—nearly matching the larger Jan-nano-128k—demonstrating its efficiency and robustness. Versus Qwen 3 1.7B Thinking, Jan-v1-edge shows a slight degradation on instruction-following and CreativeWriting, while remaining comparable or better on EQBench and recency QA. Jan-v1-edge is optimized for direct integration with the Jan App. Simply select the model from the Jan App interface for immediate access to its full capabilities. - Discussions: HuggingFace Community - Jan App: Discover more about the Jan App at jan.ai

NaNK
license:apache-2.0
139
36

mistral

license:apache-2.0
83
1

Jan-v3-4B-base-instruct-gguf

NaNK
license:apache-2.0
82
15

TinyLlama-1.1B-Chat-v1.0-GGUF

NaNK
base_model:TinyLlama/TinyLlama-1.1B-Chat-v1.0
77
3

trinity-v1-GGUF

NaNK
license:apache-2.0
73
10

distilabeled-hermes-2.5-mistral-7b-GGUF

NaNK
license:apache-2.0
64
1

Jan-code-4b

NaNK
license:apache-2.0
52
42

vi-translation_v0.1_20250612-230103_Qwen3-1.7B_checkpoint-500_vi-en

NaNK
46
0

stealth-rag-v1.1-GGUF

NaNK
license:apache-2.0
43
0

nitro-v1.2-e3-GGUF

NaNK
license:apache-2.0
38
0

llamacorn-1.1b-chat-GGUF

NaNK
base_model:jan-hq/LlamaCorn-1.1B-Chat
36
1

nitro-v1.2-e3-GGUF-tke

NaNK
license:apache-2.0
36
0

meta-llama-3-8b-instruct-healed-GGUF

NaNK
base_model:jan-hq/Meta-Llama-3-8b-Instruct-Healed
36
0

pandora-v1-13b-GGUF

NaNK
license:apache-2.0
34
1

fused_model_test_llama3_on-GGUF

base_model:alandao/fused_model_test_llama3_on
34
0

jan-repo-v1-sft-GGUF

license:apache-2.0
31
0

pandora-v1-10.7b-GGUF

NaNK
license:apache-2.0
25
2

medicine-llm-GGUF

25
0

neural-chat-7b-v3-3-slerp-GGUF

NaNK
24
0

go-bruins-v2-GGUF

license:cc-by-nc-4.0
24
0

jan-repo-v1-dpo-low-GGUF

license:apache-2.0
24
0

stealth-finance-v4-GGUF

NaNK
license:apache-2.0
24
0

Mistral-7B-Instruct-v0.2-GGUF

NaNK
license:apache-2.0
23
2

SeaLLM-7B-Chat-GGUF

NaNK
license:apache-2.0
22
1

openhermes-2.5-neuralchat-v3-2-7b-GGUF

NaNK
22
0

mysticoder-v1-GGUF

NaNK
license:apache-2.0
22
0

TenyxChat-7B-v1-GGUF

NaNK
22
0

stealth-v1.3-GGUF

NaNK
license:apache-2.0
21
1

vistral-7b-chat-dpo-GGUF

NaNK
license:apache-2.0
20
3

metamath-cybertron-starling-GGUF

license:apache-2.0
20
2

stealth-v1.2-GGUF

license:apache-2.0
20
1

tinyllama-bamboo-v1.0-GGUF

NaNK
base_model:jan-hq/TinyLlama-Bamboo-v1.0
20
0

trinity-v1.2-GGUF

19
6

tulpar-7b-v2-GGUF

NaNK
license:apache-2.0
19
0

nitro-v1-e1-GGUF

NaNK
license:apache-2.0
19
0

openhermes-2.5-neural-chat-v3-3-slerp-GGUF

license:apache-2.0
18
2

finance-llm-GGUF

18
2

neuraltrix-7b-dpo-GGUF

NaNK
license:apache-2.0
18
2

llamacorn-1.1b-GGUF

NaNK
base_model:jan-hq/LlamaCorn-1.1B
18
1

jan-repo-v1-dpo-high-GGUF

license:apache-2.0
18
0

jan-repo-dpo-e30-jan_v1.1-high-GGUF

license:apache-2.0
18
0

stealth-finance-v1-GGUF

NaNK
license:apache-2.0
17
1

chupacabra-7b-v2.02-GGUF

NaNK
17
0

v1olet_marcoroni-go-bruins-merge-7b-GGUF

NaNK
license:apache-2.0
17
0

laser-dolphin-mixtral-2x7b-dpo-GGUF

NaNK
license:apache-2.0
17
0

stealth-finance-v3-GGUF

NaNK
17
0

vi-translation_v0.1_20250612-230103_Qwen3-1.7B_checkpoint-500_en-vi

NaNK
17
0

qwen3-1.7b-jan-v1-lora

- Developed by: janhq - License: apache-2.0 - Finetuned from model : unsloth/Qwen3-1.7B This qwen3 model was trained 2x faster with Unsloth and Huggingface's TRL library.

NaNK
license:apache-2.0
16
1

merged-agi-7b-GGUF

NaNK
license:apache-2.0
16
0

jan-repo-v1-dpo-high-new-GGUF

license:apache-2.0
16
0

jan-repo-v1.1-sft-GGUF

license:apache-2.0
16
0

jan-repo-v1-dpo-e30-jan_v1-cont-GGUF

license:apache-2.0
16
0

Qwen3-14B-v0.1-deepresearch-100-step-gguf

NaNK
15
3

Mistral-7B-Instruct-v0.1-GGUF

NaNK
license:apache-2.0
15
1

stealth-rag-v1-e1-GGUF

NaNK
license:apache-2.0
15
1

stealth-v1.1-GGUF

NaNK
license:apache-2.0
15
0

jan-repo-v1-dpo-high-e30-GGUF

NaNK
license:apache-2.0
15
0

yi-1.5

NaNK
license:apache-2.0
15
0

marcoroni-7b-v2-GGUF

NaNK
license:apache-2.0
14
0

marcoroni-7b-v3-GGUF

NaNK
license:apache-2.0
14
0

OpenHermes-2.5-neural-chat-v3-2-Slerp-GGUF

license:apache-2.0
14
0

stealth-v1-math-sft-GGUF

license:apache-2.0
14
0

nitro-v1.1-e3-GGUF

NaNK
license:apache-2.0
14
0

tinyllama-bamboo-v1.5-GGUF

NaNK
base_model:jan-hq/TinyLlama-Bamboo-v1.5
14
0

stealth-finance-v1-e1-GGUF

NaNK
license:apache-2.0
14
0

mistral-7b-instruct-v0.2-sliced-24-layer-GGUF

NaNK
14
0

Solar-10.7B-SLERP-GGUF

NaNK
license:apache-2.0
13
4

nitro-v1-e3-GGUF

NaNK
license:apache-2.0
13
0

Jan-v2-VL-high

license:apache-2.0
12
32

Jan-v2-VL-med

license:apache-2.0
12
9

Jan-v2-VL-max-Instruct-FP8

NaNK
license:apache-2.0
12
6

trinity-v1.1-GGUF

NaNK
license:apache-2.0
12
3

komodo-7b-chat-GGUF

NaNK
license:apache-2.0
12
1

cerebrum-1.0-7b-GGUF

NaNK
license:apache-2.0
12
0

solar-10.7b-instruct-v1.0-GGUF

NaNK
license:cc-by-nc-4.0
11
1

tinyjensen-1.1b-chat-GGUF

NaNK
license:apache-2.0
11
0

Mistral-7B-Instruct-v0.2-SLERP-GGUF

NaNK
license:apache-2.0
10
6

mistral-7b-instruct-v0.2-dare-GGUF

NaNK
license:apache-2.0
10
0

250413-llama-3.1-8b-instruct-grpo-01-no-retry-650-gguf

NaNK
llama
10
0

tinyjensen-1.1b-GGUF

NaNK
license:apache-2.0
9
1

notus-7b-v1-GGUF

NaNK
license:apache-2.0
7
0

Jan-v2-VL-low-gguf

license:apache-2.0
5
1

qwen3-4b-v0.3-deepresearch-no-think-gguf

NaNK
license:apache-2.0
5
0

noname-thinking-v0.2-step-100-0807

5
0

noname-thinking-v0.2-step-200

4
0

tinyllama

license:apache-2.0
2
1

250409-llama-3.2-3b-instruct-grpo-01-no-retry-350-lora-gguf

NaNK
llama
2
0

Phoenix-v1-8x7B

NaNK
license:apache-2.0
1
3

llama3-s-instruct-v0.3-checkpoint-7000-phase-3-exllama2

llama
1
2

Qwen3.5-35B-A3B-GGUF

NaNK
license:apache-2.0
0
1

llama3-s-instruct-v0.4-vllm-fp8

llama
0
1

250404-llama-3.2-3b-instruct-grpo-01

NaNK
0
1

demo-deep-research-model

0
1

qwen3-4b-v0.3-deepresearch-100-step-gguf

NaNK
license:apache-2.0
0
1