crumb

85 models • 1 total models in database
Sort by:

nano-mistral

license:apache-2.0
3,481
9

gpt2023

license:mit
733
18

apricot-wildflower-20

license:apache-2.0
700
2

shrink-v1

NaNK
llama
236
3

bloom-560m-RLHF-SD2-prompter-aesthetic

85
20

alpha-wolf-dreambooth

36
1

doc2desc_3b_gguf

This is Qwen/Qwen2.5-3B tuned with the following format, on a mix of handwritten and Deepseek-V3 generated descriptions (few-shot w/ handwritten descriptions) for texts from https://textfiles.com, to make sure it has the ability to label unsafe content. This is being used to generate heaps of description/document pairs for training another model to do the reverse, for automatically generating documents to create control-vectors from. | Position | Delimiter | | --- | --- | | before user | `[[DOCUMENT]]` | | after user | `[[/DOCUMENT]]` | | before assistant | `[[DESCRIPTION]]` | after assistant | `[[/DESCRIPTION]]` you may also want to add "[[" as a stop string.. light tune, isn't perfect 😅 The outputs are something like informal summaries, for example, on the first element from the C4 dataset here are some outputs (at temperature 0.8): (outputs): > ad to get better at making delicious BBQ by world class bbq champion from lonestar smoke rangers. > ad for BBQ class at lonestar smoke rangers by world class bbq champ tony balay; includes techniques, recipes > event ad: beginners BBQ Class Taking Place in Missoula! from world class bbs champion tony balay

NaNK
27
0

scaffold-18

18
0

bloom-560m-RLHF-SD2-prompter

12
12

eva-fusion-v2.22

10
7

icon-diffusion-v1-1

9
5

distilpythia

license:apache-2.0
7
4

Ducky-MoMoe-prototype-e4-causal

7
4

gpt-joke

6
2

Llama-p-small

license:apache-2.0
6
0

FLAN-OPT-6.7b-LoRA

NaNK
5
7

fake-gpt-j-17m

5
2

opentinystories-30m-complex

5
1

minipile-111m

5
0

gpt-j-6b-finetune-super-glue

NaNK
4
0

ColabInstruct-Z-1.1B

NaNK
4
0

opentinystories-68m-complex

4
0

pico-gpt-j-6.7m

3
2

FLAN-OPT-1.3b-LoRA

NaNK
3
1

FLAN-OPT-2.7b-LoRA

NaNK
3
0

bespoke-gpt-124m

3
0

test-00-switchllama-i3b-f10b-e4-init

NaNK
switchllama
3
0

mixtral-e8-nano-1gt-test

3
0

GLORT2

3
0

utf8-gelu-dec-8.5M-10KB-ctx-3GB

llama
3
0

distilpythia-cl

license:apache-2.0
2
1

opentinystories-30m-base

license:mit
2
1

switch-base-8-arxiv-abstraction

2
0

opentinystories-68m-base

license:mit
2
0

llama2-7b-moe-text-exp2-4

NaNK
2
0

llama2-7b-moe-text-exp3-4

NaNK
2
0

test-00-qlora-wizmlpmix-c0

llama
2
0

core1-base-464m-redpajama

llama
2
0

gale-large-test

2
0

king-james-bible-gzip-16line-window

2
0

llama-d1024-slimpajama-1gt-test

llama
2
0

llama-d1536-init

llama
2
0

ParaLlama-p-small

2
0

13f189-augmented-mappings-medium-control

2
0

ptune-FLAN-OPT-6.7b

NaNK
1
2

llama2-7b-moe-text-exp1-4

NaNK
1
2

model-a-48.5m

license:apache-2.0
1
2

ptune-FLAN-OPT-2.7b

NaNK
1
1

cramped-94m-8btok

NaNK
license:apache-2.0
1
1

askmistral-2-15-111m

1
1

llama2-7b-shard-bf16

NaNK
llama
1
0

Ducky-MoMoe-prototype-e4-ul2

1
0

test-00-qlora-wizmlpmix-c1

llama
1
0

test-00-qlora-wizmlpmix-c2

llama
1
0

core1-base-464m-c4

llama
1
0

44m-textbook

license:apache-2.0
1
0

d1536-250MT-full

license:apache-2.0
1
0

25m-special

1
0

qrstudy-410m-8-1

NaNK
1
0

qrstudy-410m-16-1

NaNK
1
0

qrstudy-410m-64-1

NaNK
1
0

qrstudy-gpt2-4-8

NaNK
1
0

qrstudy-gpt2-8-16

NaNK
1
0

qrstudy-gpt2-16-32

NaNK
1
0

king-james-bible-gzip-8line-window

1
0

king-james-bible-gzip-64line-window

1
0

shrink-init

llama
1
0

ParaLlama-p-micro

1
0

gpt2-medium-eb49cc

1
0

160m-plus-sauce

1
0

Instruct-GPT-J

0
26

midjourney-textual-inversions

license:mit
0
19

icon-diffusion-ckpt

0
4

essence-3b-v2

NaNK
license:cc-by-sa-4.0
0
3

Gale-medium-init

0
3

genshin-stable-inversion

0
2

eva-model-ckpt

0
2

92d52f-ame-full-7B

NaNK
0
2

gpt-j-6b-shakespeare

NaNK
0
1

dalle-paint

license:mit
0
1

aurora-1.0

0
1

GeoV-Instruct-LoRA

0
1

CGPT-124m

0
1

horizon-pythia-ft-1b

NaNK
0
1

32M-32GT-SlimPajama

NaNK
license:apache-2.0
0
1