Felladrin

298 models • 2 total models in database
Sort by:

gguf-smollm-360M-instruct-add-basics

14,200
1

gguf-MXFP4-gpt-oss-20b-Derestricted

NaNK
unlimited
5,082
5

Llama-68M-Chat-v1

llama
931
13

gguf-jina-reranker-v1-tiny-en

Model creator: Jina AI Original model: jina-reranker-v1-tiny-en GGUF quantization: based on llama.cpp release f4d2b This model is designed for blazing-fast reranking while maintaining competitive performance. What's more, it leverages the power of our JinaBERT model as its foundation. `JinaBERT` itself is a unique variant of the BERT architecture that supports the symmetric bidirectional variant of ALiBi. This allows `jina-reranker-v1-tiny-en` to process significantly longer sequences of text compared to other reranking models, up to an impressive 8,192 tokens. To achieve the remarkable speed, the `jina-reranker-v1-tiny-en` employ a technique called knowledge distillation. Here, a complex, but slower, model (like our original jina-reranker-v1-base-en) acts as a teacher, condensing its knowledge into a smaller, faster student model. This student retains most of the teacher's knowledge, allowing it to deliver similar accuracy in a fraction of the time. Here's a breakdown of the reranker models we provide: | Model Name | Layers | Hidden Size | Parameters (Millions) | | ------------------------------------------------------------------------------------ | ------ | ----------- | --------------------- | | jina-reranker-v1-base-en | 12 | 768 | 137.0 | | jina-reranker-v1-turbo-en | 6 | 384 | 37.8 | | jina-reranker-v1-tiny-en | 4 | 384 | 33.0 | > Currently, the `jina-reranker-v1-base-en` model is not available on Hugging Face. You can access it via the Jina AI Reranker API. As you can see, the `jina-reranker-v1-turbo-en` offers a balanced approach with 6 layers and 37.8 million parameters. This translates to fast search and reranking while preserving a high degree of accuracy. The `jina-reranker-v1-tiny-en` prioritizes speed even further, achieving the fastest inference speeds with its 4-layer, 33.0 million parameter architecture. This makes it ideal for scenarios where absolute top accuracy is less crucial. 1. The easiest way to starting using `jina-reranker-v1-tiny-en` is to use Jina AI's Reranker API. 2. Alternatively, you can use the latest version of the `sentence-transformers>=0.27.0` library. You can install it via pip: Then, you can use the following code to interact with the model: 3. You can also use the `transformers` library to interact with the model programmatically. 4. You can also use the `transformers.js` library to run the model directly in JavaScript (in-browser, Node.js, Deno, etc.)! If you haven't already, you can install the Transformers.js JavaScript library from NPM using: Then, you can use the following code to interact with the model: That's it! You can now use the `jina-reranker-v1-tiny-en` model in your projects. We evaluated Jina Reranker on 3 key benchmarks to ensure top-tier performance and search relevance. | Model Name | NDCG@10 (17 BEIR datasets) | NDCG@10 (5 LoCo datasets) | Hit Rate (LlamaIndex RAG) | | ------------------------------------------ | -------------------------- | ------------------------- | ------------------------- | | `jina-reranker-v1-base-en` | 52.45 | 87.31 | 85.53 | | `jina-reranker-v1-turbo-en` | 49.60 | 69.21 | 85.13 | | `jina-reranker-v1-tiny-en` (you are here) | 48.54 | 70.29 | 85.00 | | `mxbai-rerank-base-v1` | 49.19 | - | 82.50 | | `mxbai-rerank-xsmall-v1` | 48.80 | - | 83.69 | | `ms-marco-MiniLM-L-6-v2` | 48.64 | - | 82.63 | | `ms-marco-MiniLM-L-4-v2` | 47.81 | - | 83.82 | | `bge-reranker-base` | 47.89 | - | 83.03 | - `NDCG@10` is a measure of ranking quality, with higher scores indicating better search results. `Hit Rate` measures the percentage of relevant documents that appear in the top 10 search results. - The results of LoCo datasets on other models are not available since they do not support long documents more than 512 tokens. For more details, please refer to our benchmarking sheets. Join our Discord community and chat with other community members about ideas.

license:apache-2.0
710
2

TinyMistral-248M-Chat-v4

license:apache-2.0
401
31

Minueza-32M-Base

license:apache-2.0
391
19

gguf-sharded-Qwen2-0.5B-Instruct

NaNK
license:apache-2.0
361
0

gguf-Qwen1.5-0.5B-Chat

NaNK
262
2

gguf-flan-t5-small

license:apache-2.0
255
3

gguf-gemma-2b-orpo

NaNK
186
1

gguf-flan-t5-large

license:apache-2.0
128
5

gguf-Aira-2-355M

license:apache-2.0
110
0

gguf-Q2_K_S-Mixed-AutoRound-MiniMax-M2.1

NaNK
107
5

gguf-TinyMistral-248M-Chat-v2

NaNK
license:apache-2.0
107
0

gguf-pythia-1.4b-sft-full

NaNK
license:apache-2.0
106
0

gguf-multi-qa-MiniLM-L6-cos-v1

NaNK
101
0

gguf-Smol-Llama-101M-Chat-v1

NaNK
base_model:Felladrin/Smol-Llama-101M-Chat-v1
92
0

gguf-Qwen2-0.5B-Instruct

NaNK
license:apache-2.0
91
0

gguf-Q5_K_M-Qwen2.5-0.5B-Instruct

NaNK
llama-cpp
89
0

gguf-MobileLLaMA-1.4B-Chat

NaNK
base_model:mtgv/MobileLLaMA-1.4B-Chat
87
0

gguf-Phi-3-mini-4k-instruct

license:mit
86
0

gguf-openhermes-tinyllama-sft-qlora

base_model:Ritvik19/openhermes-tinyllama-sft-qlora
86
0

gguf-LaMini-Flan-T5-248M

license:cc-by-nc-4.0
85
0

gguf-flan-t5-base

license:apache-2.0
81
0

gguf-sharded-LaMini-Flan-T5-783M

license:cc-by-nc-4.0
80
0

gguf-q5_k_m-granite-3.0-2b-instruct

NaNK
79
0

gguf-WizardVicuna-pythia-410m-deduped

78
2

gguf-sharded-Aira-2-355M

license:apache-2.0
74
0

gguf-sharded-WizardVicuna-pythia-410m-deduped

70
1

gguf-Qwen1.5-0.5B-Chat_llamafy

NaNK
base_model:Minami-su/Qwen1.5-0.5B-Chat_llamafy
70
0

gguf-sharded-Qwen2-1.5B-Instruct

NaNK
license:apache-2.0
67
0

gguf-sharded-Qwen1.5-0.5B-Chat_llamafy

NaNK
base_model:Minami-su/Qwen1.5-0.5B-Chat_llamafy
67
0

gguf-Qwen2-1.5B-Instruct

NaNK
license:apache-2.0
65
0

gguf-SmolLM-135M-Instruct

65
0

gguf-Qwen2-0.5B-Instruct-llamafy

NaNK
base_model:Minami-su/Qwen2-0.5B-Instruct-llamafy
64
0

gguf-sharded-Qwen2-0.5B-Instruct-llamafy

NaNK
base_model:Minami-su/Qwen2-0.5B-Instruct-llamafy
62
0

gguf-sharded-gemma-2b-orpo

NaNK
61
1

gguf-sharded-UD-Q4_K_XL-Qwen3-0.6B

NaNK
61
0

gguf-sharded-Phi-3-mini-4k-instruct

license:mit
59
0

Llama-160M-Chat-v1

Language: en License: apache-2.0

llama
58
20

gguf-zephyr-220m-dpo-full

56
1

gguf-Llama-160M-Chat-v1

NaNK
base_model:Felladrin/Llama-160M-Chat-v1
55
2

gguf-h2o-danube3-500m-chat

55
1

gguf-spin_gpt2_medium_alpaca_e2

license:mit
55
0

Smol-Llama-101M-Chat-v1

llama
53
9

gguf-TinyMistral-248M-SFT-v4

NaNK
license:apache-2.0
53
1

gguf-LaMini-Flan-T5-77M

license:cc-by-nc-4.0
53
0

gguf-Q8_0-bge-reranker-v2-m3

NaNK
llama-cpp
51
2

gguf-gemma-2-2b-it-abliterated

NaNK
50
0

gguf-sharded-TinyMistral-248M-Chat-v2

NaNK
license:apache-2.0
49
0

gguf-1.5-Pints-2K-v0.1

NaNK
49
0

gguf-sharded-openhermes-1b-olmo-sft-qlora

NaNK
license:apache-2.0
45
0

gguf-internlm2-chat-1_8b

NaNK
44
0

gguf-sharded-internlm2-chat-1_8b

NaNK
44
0

gguf-t5-base-grammar-correction

43
2

gguf-Pythia-Chat-Base-7B

NaNK
license:apache-2.0
43
0

gguf-MicroLlama

base_model:keeeeenw/MicroLlama
43
0

gguf-TinySolar-248m-4k-code-instruct

license:apache-2.0
43
0

gguf-Q5_K_M-Fox-1-1.6B-Instruct-v0.1

NaNK
llama-cpp
43
0

gguf-sharded-flan-t5-large

license:apache-2.0
42
0

gguf-openhermes-1b-olmo-sft-qlora

NaNK
license:apache-2.0
41
1

gguf-zephyr-1b-olmo-sft-qlora

NaNK
license:apache-2.0
40
0

gguf-sharded-TinySolar-248m-4k-code-instruct

license:apache-2.0
39
0

gguf-Lite-Mistral-150M-v2-Instruct

39
0

gguf-prem-1B-chat

NaNK
license:apache-2.0
38
1

gguf-q8_0-madlad400-3b-mt

NaNK
38
1

gguf-sharded-Llama-160M-Chat-v1

NaNK
base_model:Felladrin/Llama-160M-Chat-v1
37
0

gguf-sharded-h2o-danube2-1.8b-chat

NaNK
license:apache-2.0
36
1

gguf-llama-160m

base_model:JackFram/llama-160m
36
0

gguf-Lite-Oute-1-65M-Instruct

36
0

gguf-Q8_0-Qwen2.5-0.5B-Instruct

NaNK
36
0

gguf-sharded-Qwen1.5-0.5B-Chat

NaNK
35
0

gguf-Q5_K_M-smollm-360M-instruct-add-basics

llama-cpp
35
0

gguf-sharded-falcon-mamba-7b-instruct

NaNK
35
0

gguf-Hare-1.1B-Chat

NaNK
34
0

gguf-OLMoE-1B-7B-0924-Instruct

NaNK
34
0

gguf-q5_k_m-madlad400-3b-mt

NaNK
33
1

gguf-sharded-prem-1B-chat

NaNK
license:apache-2.0
32
0

gguf-h2o-danube2-1.8b-chat

NaNK
license:apache-2.0
31
1

gguf-NuExtract-tiny

30
2

candle-quantized-LaMini-Flan-T5-248M

license:cc-by-nc-4.0
30
0

gguf-flan-t5-base-instruct-dolly_hhrlhf

license:cc-by-sa-3.0
30
0

gguf-Q4_K_M-Yi-1.5-6B-Chat

NaNK
llama-cpp
30
0

gguf-TinyMistral-248M-Chat-v1

NaNK
license:apache-2.0
29
0

gguf-smol_llama-220M-openhermes

base_model:BEE-spoke-data/smol_llama-220M-openhermes
29
0

gguf-sharded-MobileLLaMA-1.4B-Chat

NaNK
base_model:mtgv/MobileLLaMA-1.4B-Chat
28
0

gguf-DopeyTinyLlama-1.1B-v1

NaNK
base_model:vihangd/DopeyTinyLlama-1.1B-v1
27
0

gguf-flan-alpaca-base

license:apache-2.0
27
0

gguf-gpt2-chatbot

license:apache-2.0
26
0

gguf-sharded-zephyr-1b-olmo-sft-qlora

NaNK
license:apache-2.0
26
0

gguf-Sheared-Pythia-160m-Platypus

license:cc-by-nc-sa-4.0
26
0

gguf-t5-address-standardizer

25
0

gguf-sharded-Q5_K_L-Llama-3.2-3B-Instruct

NaNK
base_model:bartowski/Llama-3.2-3B-Instruct-GGUF
25
0

gguf-Q8_0-SmolLM2-360M-Instruct

GGUF version of HuggingFaceTB/SmolLM2-360M-Instruct.

25
0

gguf-Pythia-31M-Chat-v1

NaNK
license:apache-2.0
24
0

gguf-pythia-3b-deduped-sft

NaNK
license:apache-2.0
24
0

gguf-MaxMini-Instruct-248M

license:mit
24
0

gguf-SmolLM-360M-Instruct

24
0

gguf-sharded-Qwen2-1.5B-Instruct-imat

NaNK
llama-cpp
24
0

gguf-Q4_K_S-MiniCPM4-0.5B-QAT-Int4-unquantized

Felladrin/MiniCPM4-0.5B-QAT-Int4-unquantized-Q4KS-GGUF This model was converted to GGUF format from `openbmb/MiniCPM4-0.5B-QAT-Int4-unquantized` using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model. Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well. Step 2: Move into the llama.cpp folder and build it with `LLAMACURL=1` flag along with other hardware-specific flags (for ex: LLAMACUDA=1 for Nvidia GPUs on Linux).

NaNK
llama-cpp
24
0

Pythia-31M-Chat-v1

license:apache-2.0
23
8

gguf-TinyMistral-248M-v2.5-Instruct-orpo

license:apache-2.0
23
0

gguf-q5_k_l-imat-arcee-lite

license:apache-2.0
23
0

gguf-Q3_K_XL-falcon-mamba-7b

NaNK
23
0

gguf-vicuna-68m

license:apache-2.0
22
0

gguf-TinyLlama-1.1B-1T-OpenOrca

NaNK
base_model:jeff31415/TinyLlama-1.1B-1T-OpenOrca
22
0

gguf-sharded-spin_gpt2_medium_alpaca_e2

license:mit
22
0

gguf-IPythia-410m

22
0

gguf-LaMini-Flan-T5-783M

license:cc-by-nc-4.0
22
0

gguf-TinySolar-248m-4k

license:apache-2.0
22
0

gguf-sharded-Q5_K_L-Llama-3.2-1B-Instruct

NaNK
base_model:bartowski/Llama-3.2-1B-Instruct-GGUF
22
0

gguf-Q5_K_M-NanoLM-1B-Instruct-v2

NaNK
llama-cpp
21
1

gguf-sharded-pythia-3b-deduped-sft

NaNK
license:apache-2.0
21
0

gguf-Aira-2-124M

license:apache-2.0
21
0

gguf-Aira-2-124M-DPO

license:apache-2.0
21
0

gguf-MiniMA-2-1B

NaNK
21
0

gguf-q5_k_m-phi-3.5-mini-instruct

llama-cpp
21
0

gguf-sharded-openhermes-tinyllama-sft-qlora

base_model:Ritvik19/openhermes-tinyllama-sft-qlora
20
0

gguf-sharded-Aira-2-124M-DPO

license:apache-2.0
20
0

gguf-sharded-IPythia-410m

20
0

gguf-sharded-TinyMistral-248M-v2.5-Instruct-orpo

license:apache-2.0
20
0

gguf-sharded-pythia-1.4b-sft-full

NaNK
license:apache-2.0
19
0

gguf-sharded-Aira-2-124M

license:apache-2.0
19
0

gguf-falcon-mamba-7b-instruct

NaNK
19
0

Minueza-2-96M-Instruct-Variant-10

llama
19
0

Minueza-32M-UltraChat

Language model with Apache 2.0 license.

license:apache-2.0
18
5

gguf-sharded-LaMini-Flan-T5-248M

llama-cpp
18
1

gguf-774M-03_09_2024

NaNK
license:mit
18
0

gguf-gpt2-alpaca-gpt4

NaNK
license:mit
18
0

gguf-flan-t5-small-finetuned-openai-summarize_from_feedback

18
0

gguf-sharded-gemma-2-2b-it-abliterated

NaNK
llama-cpp
18
0

gguf-Q5_K_M-Qwen3-4B-Merge-Variant-01

Felladrin/gguf-Q5KM-Qwen3-4B-Merge-Variant-01 This model was converted to GGUF format from `Felladrin/Qwen3-4B-Merge-Variant-01` using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model. Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well. Step 2: Move into the llama.cpp folder and build it with `LLAMACURL=1` flag along with other hardware-specific flags (for ex: LLAMACUDA=1 for Nvidia GPUs on Linux).

NaNK
llama-cpp
18
0

gguf-sharded-Q4_K_S-AFM-4.5B

Sharded GGUF version of bartowski/arcee-aiAFM-4.5B-GGUF.

NaNK
18
0

gguf-mamba-130m-hf

17
0

gguf-Q5_K_M-NanoLM-0.3B-Instruct-v2

NaNK
llama-cpp
17
0

Minueza-2-96M

llama
16
6

gguf-Lite-Oute-1-300M-Instruct

16
0

gguf-sharded-Q4_K_S-gemma-3-270m-it

Sharded GGUF version of bartowski/googlegemma-3-270m-it-GGUF.

16
0

gguf-q5_k_m-tinydolphin-2.8.2-1.1b-laser

NaNK
llama-cpp
15
0

gguf-Q5_K_M-MagpieLM-4B-Chat-v0.1

NaNK
llama-cpp
15
0

gguf-sharded-UD-Q4_K_XL-Phi-4-mini-reasoning

Sharded GGUF version of unsloth/Phi-4-mini-reasoning-GGUF.

15
0

gguf-sharded-Q4_K_S-MiniCPM4-0.5B-QAT

Sharded GGUF version of Felladrin/MiniCPM4-0.5B-QAT-Int4-unquantized-Q4KS-GGUF.

NaNK
15
0

gguf-Minueza-32M-Chat

license:apache-2.0
14
0

gguf-Minueza-32Mx2-Chat

license:apache-2.0
13
0

gguf-IQ3_XXS-OLMo-7B-0424-Instruct-hf

NaNK
llama-cpp
13
0

gguf-sharded-UD-Q4_K_XL-Qwen3-1.7B

NaNK
13
0

gguf-sharded-Q4_K_S-Polaris-4B-Preview

NaNK
13
0

gguf-sharded-Q4_K_S-cogito-v1-preview-llama-3B

NaNK
base_model:bartowski/deepcogito_cogito-v1-preview-llama-3B-GGUF
13
0

gguf-Tinyllama-616M-Cinder

base_model:Josephgflowers/Tinyllama-616M-Cinder
12
1

gguf-q5_k_m-imat-qwen2-0.5b-instruct

NaNK
llama-cpp
12
0

gguf-Q5_K_M-Nemotron-Mini-4B-Instruct

NaNK
llama-cpp
12
0

gguf-Q8_0-smollm-135M-instruct-v0.2

NaNK
llama-cpp
10
1

gguf-vicuna-160m

license:apache-2.0
10
0

gguf-phi-1_5

NaNK
license:mit
10
0

gguf-flan-t5-small-cnndm

10
0

gguf-Q8_0-all-MiniLM-L6-v2

NaNK
llama-cpp
10
0

gguf-Mixnueza-6x32M-MoE

license:apache-2.0
9
1

gguf-Minueza-32M-Base

license:apache-2.0
9
0

gguf-q5_k_m-h2o-danube2-1.8b-chat

NaNK
9
0

gguf-Q5_K_M-LlamaCorn-1.1B-Chat

NaNK
llama-cpp
9
0

gguf-Q8_0-Qwen2.5-Coder-1.5B-Instruct

NaNK
llama-cpp
9
0

mlx-5bit-Qwen3-4B-Merge-Variant-01

NaNK
license:apache-2.0
9
0

Qwen2-96M

license:apache-2.0
8
2

gguf-Minueza-32M-UltraChat

license:apache-2.0
8
1

gguf-sharded-TinyLlama-1.1B-1T-OpenOrca

NaNK
base_model:jeff31415/TinyLlama-1.1B-1T-OpenOrca
8
0

gguf-sharded-h2o-danube3-500m-chat

license:apache-2.0
8
0

gguf-Q3_K_L-Yi-1.5-6B-Chat

NaNK
license:apache-2.0
8
0

gguf-sharded-Q4_K_S-DeepSeek-R1-Distill-Qwen-1.5B

NaNK
8
0

gguf-sharded-Q4_K_S-LFM2-350M

8
0

gguf-sharded-Q4_K_S-LFM2-700M

8
0

gguf-sharded-OLMo-7B-Instruct

NaNK
license:apache-2.0
7
0

gguf-Q5_K_L-Nemotron-Mini-4B-Instruct

NaNK
7
0

gguf-Q5_K_M-SmolLM2-1.7B-Instruct

GGUF version of HuggingFaceTB/SmolLM2-1.7B-Instruct.

NaNK
7
0

gguf-sharded-Q4_K_S-SmolLM3-3B

Sharded GGUF version of bartowski/HuggingFaceTBSmolLM3-3B-GGUF.

NaNK
7
0

gguf-sharded-Q4_K_S-granite-3.3-2b-instruct

NaNK
7
0

Minueza-32M-Chat

license:apache-2.0
6
8

gguf-1.5-Pints-16K-v0.1

NaNK
6
1

onnx-gpt2-medium-chat

license:mit
6
0

onnx-gpt2-conversational-retrain

license:mit
6
0

gguf-sharded-vicuna-160m

license:apache-2.0
6
0

gguf-Q5_K_M-Phi-1_5-Instruct-v0.1

NaNK
llama-cpp
6
0

gguf-sharded-Q4_K_S-OLMoE-1B-7B-0924-Instruct

NaNK
6
0

gguf-sharded-Q4_K_S-LFM2-1.2B

NaNK
6
0

Minueza-2-96M-Instruct-Variant-02

llama
5
1

gguf-sharded-Q3_K_L-OLMoE-1B-7B-0924-Instruct

NaNK
5
0

gguf-sharded-Q4_K_S-Llama-3.1-Nemotron-Nano-4B-v1.1

Sharded GGUF version of bartowski/nvidiaLlama-3.1-Nemotron-Nano-4B-v1.1-GGUF.

NaNK
base_model:bartowski/nvidia_Llama-3.1-Nemotron-Nano-4B-v1.1-GGUF
5
0

gguf-sharded-stablelm-2-1_6b-chat

NaNK
4
0

gguf-sharded-phi-2-orange-v2

NaNK
license:mit
4
0

gguf-q5_k_m-h2o-danube3-500m-chat

llama-cpp
4
0

gguf-Q2_K_L-Llama-3.1-SuperNova-Lite

base_model:arcee-ai/Llama-3.1-SuperNova-Lite
4
0

gguf-sharded-Q5_K_L-granite-3.0-3b-a800m-instruct

Sharded GGUF version of bartowski/granite-3.0-3b-a800m-instruct-GGUF.

NaNK
4
0

gguf-sharded-Q8_0-Qwen2.5-Coder-0.5B-Instruct

NaNK
4
0

gguf-Q4_0-Qwen2.5-Coder-32B-Instruct-abliterated

NaNK
llama-cpp
4
0

gguf-sharded-UD-Q4_K_XL-OLMo-2-0425-1B-Instruct

NaNK
4
0

Qwen3-4B-Merge-Variant-01

NaNK
license:apache-2.0
4
0

llama2_xs_460M_experimental_evol_instruct

llama
3
5

onnx-megatron-gpt2-345m-evol_instruct_v2

license:cc-by-nc-4.0
3
0

onnx-Pythia-31M-Chat-v1

NaNK
license:apache-2.0
3
0

onnx-Minueza-32M-Chat

license:apache-2.0
3
0

gguf-mamba-370m-hf

llama-cpp
3
0

gguf-stablelm-2-1_6b-chat

NaNK
3
0

gguf-sharded-zephyr-220m-dpo-full

license:apache-2.0
3
0

gguf-Q4_K_S-OLMo-7B-0424-Instruct-hf

NaNK
llama-cpp
3
0

gguf-Q5_K_M-NanoLM-70M-Instruct-v1

NaNK
llama-cpp
3
0

gguf-sharded-Q5_K_L-Replete-LLM-V2.5-Qwen-3b

NaNK
3
0

gguf-sharded-Q5_K_L-Replete-LLM-V2.5-Qwen-0.5b

NaNK
3
0

gguf-Q8_0-SmolLM2-135M-Instruct

GGUF version of HuggingFaceTB/SmolLM2-135M-Instruct.

3
0

gguf-Q4_K_M-MiniCPM3-4B

NaNK
3
0

gguf-Q8_0-LaMini-Flan-T5-248M

llama-cpp
3
0

gguf-sharded-Q4_K_S-gemma-3-4b-it

NaNK
3
0

Minueza-2-96M-Instruct-Variant-03

llama
3
0

gguf-sharded-Q4_K_S-Apriel-5B-Instruct-llamafied

NaNK
base_model:mrfakename/Apriel-5B-Instruct-llamafied
3
0

gguf-sharded-Q4_K_S-h2o-danube3.1-4b-chat

NaNK
3
0

gguf-sharded-Q4_K_S-gemma-3n-E2B-it

Sharded GGUF version of bartowski/googlegemma-3n-E2B-it-GGUF.

NaNK
3
0

gguf-sharded-Q4_K_S-Falcon-H1-0.5B-Instruct

Sharded GGUF version of mradermacher/Falcon-H1-0.5B-Instruct-i1-GGUF.

NaNK
3
0

Minueza-32Mx2-Chat

license:apache-2.0
2
2

Sheared-Pythia-160m-Platypus

license:cc-by-nc-sa-4.0
2
1

onnx-gpt2-large-conversational-retrain

license:mit
2
0

onnx-GPT2-Medium-Alpaca-355m

license:mit
2
0

onnx-llama2_xs_460M_experimental_evol_instruct

llama
2
0

onnx-gpt2-alpaca

license:mit
2
0

gguf-sharded-Phi-3-mini-4k-instruct-iMat

license:mit
2
0

gguf-sharded-smashed-WizardLM-2-7B

NaNK
2
0

gguf-sharded-Mistral-7B-OpenOrca

NaNK
license:apache-2.0
2
0

gguf-sharded-wavecoder-ultra-6.7b

NaNK
license:mit
2
0

mlc-q4f16-Phi-3.5-mini-instruct

2
0

gguf-Q5_K_M-TinyJensen-1.1B-Chat

NaNK
llama-cpp
2
0

gguf-Q5_K_M-fastchat-t5-3b-v1.0

NaNK
2
0

gguf-Q5_K_M-Sheared-LLaMA-1.3B-ShareGPT

NaNK
llama-cpp
2
0

gguf-Q5_K_M-OLMo-1B-SFT-hf

NaNK
llama-cpp
2
0

gguf-sharded-q3_k_m-jais-adapted-7b-chat

Sharded GGUF version of QuantFactory/jais-adapted-7b-chat-GGUF.

NaNK
2
0

gguf-q8_0-h2o-danube3-500m-chat

2
0

gguf-sharded-Q4_K_S-Qwen2.5-0.5B-Instruct

NaNK
2
0

gguf-sharded-q5_k_m-internlm2_5-1_8b-chat

Sharded GGUF version of internlm/internlm25-18b-chat-gguf.

NaNK
2
0

gguf-Q4_0-Qwen2.5-Coder-32B-Instruct

NaNK
llama-cpp
2
0

gguf-sharded-Q4_K_S-SmolLM2-135M-Instruct

2
0

gguf-sharded-Q4_K_S-TAID-LLM-1.5B

NaNK
2
0

gguf-sharded-Q3_K_M-OLMoE-1B-7B-0125-Instruct

NaNK
2
0

Minueza-2-96M-Instruct-Variant-06

llama
2
0

Minueza-2-96M-Instruct-Variant-08

llama
2
0

onnx-TinyMistral-248M-v2

NaNK
license:apache-2.0
1
3

LaMini-Neo-125M-Evol-Instruct

license:cc-by-nc-4.0
1
1

onnx-bloomz-560m-sft-chat

1
1

llama2_xs_460M_experimental_platypus

llama
1
1

onnx-flan-alpaca-base

license:apache-2.0
1
0

onnx-Cerebras-GPT-111M-instruction

license:apache-2.0
1
0

onnx-flan-t5-base-samsum

license:apache-2.0
1
0

onnx-Evol-Orca-LaMini-flan-t5-small

1
0

onnx-Smol-Llama-101M-Chat-v1

NaNK
llama
1
0

onnx-tinyllama-15M

llama
1
0

onnx-tinyllama-42M

llama
1
0

onnx-Gerbil-A-32m

license:apache-2.0
1
0

Minueza-32M-Deita

NaNK
license:apache-2.0
1
0

onnx-TinyMistral-248M-Chat-v1

NaNK
license:apache-2.0
1
0

gguf-TinyLlama-1.1B-Chat-v1.0

NaNK
llama-cpp
1
0

mlc-q4f16_1-gemma-2-2b-it

NaNK
1
0

gguf-sharded-Q3_K_XL-OLMoE-1B-7B-0924-Instruct

NaNK
1
0

gguf-Q5_K_M-TinyLlama-1.1B-Chat-v1.0

NaNK
llama-cpp
1
0

gguf-sharded-Q5_K_M-TinyLlama-1.1B-Chat-v1.0

NaNK
base_model:Felladrin/gguf-Q5_K_M-TinyLlama-1.1B-Chat-v1.0
1
0

gguf-sharded-Q5_K_M-LlamaCorn-1.1B-Chat

NaNK
base_model:Felladrin/gguf-Q5_K_M-LlamaCorn-1.1B-Chat
1
0

gguf-sharded-F16-1.5-Pints-2K-v0.1

1
0

gguf-sharded-BF16-1.5-Pints-16K-v0.1

1
0

gguf-sharded-Q5_K-1.5-Pints-2K-v0.1

NaNK
1
0

gguf-sharded-q5_k_l-granite-3.0-1b-a400m-instruct

NaNK
1
0

gguf-Q5_K_L-AMD-OLMo-1B-SFT-DPO

Sharded GGUF version of bartowski/AMD-OLMo-1B-SFT-DPO-GGUF.

NaNK
1
0

gguf-sharded-Q5_K_L-h2o-danube3-500m-chat

1
0

gguf-sharded-Q5_K_M-EXAONE-3.5-2.4B-Instruct

NaNK
1
0

gguf-sharded-Q4_K_S-granite-3.1-1b-a400m-instruct

NaNK
1
0

gguf-sharded-Q4_K_S-SmolLM2-360M-Instruct

Sharded GGUF version of bartowski/SmolLM2-360M-Instruct-GGUF.

1
0

gguf-sharded-Q4_K_S-AMD-OLMo-1B-SFT-DPO

NaNK
1
0

gguf-sharded-Q4_K_S-h2o-danube3-500m-chat

1
0

gguf-sharded-Q4_K_S-MiniCPM3-4B

NaNK
1
0

gguf-sharded-Q4_K_S-Phi-3.5-mini-instruct

1
0

gguf-sharded-Q4_K_S-Falcon3-1B-Instruct

NaNK
1
0

gguf-sharded-Q4_K_S-granite-3.1-3b-a800m-instruct

NaNK
1
0

gguf-sharded-Q4_K_S-pythia-1.4b-sft-full

Sharded GGUF version of Felladrin/gguf-pythia-1.4b-sft-full.

NaNK
1
0

gguf-Q4_K_S-1.5-Pints-16K-v0.1

NaNK
llama-cpp
1
0

gguf-sharded-Q4_K_S-internlm2_5-1_8b-chat

NaNK
1
0

gguf-sharded-Q4_K_S-EXAONE-3.5-2.4B-Instruct

NaNK
1
0

gguf-sharded-Q4_K_S-MagpieLM-4B-Chat-v0.1

NaNK
1
0

gguf-sharded-Q4_K_S-Nemotron-Mini-4B-Instruct

NaNK
1
0

gguf-sharded-Q4_K_S-stablelm-2-zephyr-1.6b

Sharded GGUF version of second-state/stablelm-2-zephyr-1.6b-GGUF.

NaNK
1
0

gguf-sharded-Q4_K_S-Megrez-3B-Instruct

NaNK
1
0

gguf-sharded-Q4_K_S-AceInstruct-1.5B

NaNK
1
0

gguf-sharded-Q4_K_S-gemma-3-1b-it

NaNK
1
0

Minueza-2-96M-Instruct-Variant-07

llama
1
0

onnx-TinyMistral-248M

license:apache-2.0
0
7

Sheared-Pythia-160m-WebGLM-QA

license:apache-2.0
0
3

mlc-chat-Mistral-7B-OpenOrca-q4f32_1

NaNK
license:apache-2.0
0
2

mlc-chat-Mistral-7B-Instruct-v0.1-q4f32_1

NaNK
license:apache-2.0
0
1

onnx-zephyr-smol_llama-100m-dpo-full

llama
0
1

onnx-LaMini-Neo-125M-Evol-Instruct

license:cc-by-nc-4.0
0
1

mlc-chat-llama-160m-q4f32_1

license:apache-2.0
0
1

onnx-Llama-160M-Chat-v1

NaNK
llama
0
1

onnx-Minueza-32M-UltraChat

license:apache-2.0
0
1

mlc-q0f16-h2o-danube3-500m-chat

license:apache-2.0
0
1

gguf-sharded-Q4_K_S-granite-3.1-2b-instruct

NaNK
0
1

Minueza-2-96M-Instruct-Variant-01

llama
0
1