diffusers

58 models • 1 total models in database
Sort by:

stable-diffusion-xl-1.0-inpainting-0.1

--- license: openrail++ base_model: stabilityai/stable-diffusion-xl-base-1.0 tags: - stable-diffusion-xl - stable-diffusion-xl-diffusers - text-to-image - diffusers - inpainting inference: false ---

337,839
345

controlnet-depth-sdxl-1.0

NaNK
17,409
190

controlnet-canny-sdxl-1.0

These are controlnet weights trained on stabilityai/stable-diffusion-xl-base-1.0 with canny conditioning. You can find some example images in the following. prompt: a couple watching a romantic sunset, 4k photo prompt: a woman, close up, detailed, beautiful, street photography, photorealistic, detailed, Kodak ektar 100, natural, candid shot prompt: Cinematic, neoclassical table in the living room, cinematic, contour, lighting, highly detailed, winter, golden hour prompt: a tornado hitting grass field, 1980's film grain. overcast, muted colors. To more details, check out the official documentation of `StableDiffusionXLControlNetPipeline`. Our training script was built on top of the official training script that we provide here. Training data This checkpoint was first trained for 20,000 steps on laion 6a resized to a max minimum dimension of 384. It was then further trained for 20,000 steps on laion 6a resized to a max minimum dimension of 1024 and then filtered to contain only minimum 1024 images. We found the further high resolution finetuning was necessary for image quality. Batch size Data parallel with a single gpu batch size of 8 for a total batch size of 64. Hyper Parameters Constant learning rate of 1e-4 scaled by batch size for total learning rate of 64e-4

NaNK
10,092
518

sdxl-instructpix2pix-768

8,817
53

FLUX.2-dev-bnb-4bit

NaNK
5,289
14

controlnet-depth-sdxl-1.0-mid

NaNK
2,288
18

ddpm_dummy

1,576
1

controlnet-depth-sdxl-1.0-small

NaNK
1,466
19

controlnet-zoe-depth-sdxl-1.0

These are ControlNet weights trained on stabilityai/stable-diffusion-xl-base-1.0 with zoe depth conditioning. Zoe-depth is an open-source SOTA depth estimation model which produces high-quality depth maps, which are better suited for conditioning. To more details, check out the official documentation of `StableDiffusionXLControlNetPipeline`. Our training script was built on top of the official training script that we provide here. Training data and Compute The model is trained on 3M image-text pairs from LAION-Aesthetics V2. The model is trained for 700 GPU hours on 80GB A100 GPUs. Batch size Data parallel with a single gpu batch size of 8 for a total batch size of 256.

NaNK
657
39

controlnet-canny-sdxl-1.0-small

NaNK
567
20

FLUX.1-dev-bnb-4bit

NaNK
544
5

t5-nf4

403
2

FLUX.1-vae

373
4

controlnet-canny-sdxl-1.0-mid

NaNK
299
18

tiny-stable-diffusion-torch

license:apache-2.0
223
3

sd-vae-ft-mse

license:mit
202
2

FLUX.1-dev-torchao-int8

126
6

controlnet-sd-xl-0.9

122
1

FLUX.1-dev-torchao-fp8

Visual comparison of Flux-dev model outputs using BF16 and torchao float8weightonly quantization To use this quantized FLUX.1 [dev] checkpoint, you need to install the 🧨 diffusers and torchao library: After installing the required library, you can run the following script: This checkpoint was created with the following script using "black-forest-labs/FLUX.1-dev" checkpoint:

116
2

FLUX.1-dev-bnb-8bit

NaNK
99
4

HunyuanVideo-vae

45
2

qwen-image-nf4

34
0

lora-trained-xl

NaNK
33
4

lora-trained-xl-keramer-face

NaNK
31
1

sdxl-vae-fp16-fix

27
0

pix2pix-sd

18
5

FLUX.1-dev-torchao-int4

Visual comparison of Flux-dev model outputs using BF16 and torchao int4weightonly quantization To use this quantized FLUX.1 [dev] checkpoint, you need to install the 🧨 diffusers and torchao library: For now, we require this specific branch in diffusers library to fix an error when loading the model After installing the required library, you can run the following script: This checkpoint was created with the following script using "black-forest-labs/FLUX.1-dev" checkpoint:

18
1

LTX-Video-0.9.0

18
0

tools

14
10

lora-trained-xl-starbucks

NaNK
13
0

lora-trained-xl-potato-head

NaNK
12
0

ddpm-cifar10-32-demo

license:apache-2.0
10
1

t2iadapter_keypose_sd14v1

9
0

ddpm-dummy

8
0

stable-diffusion-2-1-unclip-i2i-l

5
3

shot-categorizer-v0

NaNK
license:mit
5
2

t2iadapter_depth_sd15v2

5
0

t2iadapter_sketch_sd15v2

5
0

motion-adapter-test

5
0

consistency-decoder-test

5
0

t2iadapter_canny_sd14v1

4
0

t2iadapter_canny_sd15v2

4
0

t2iadapter_depth_sd14v1

4
0

t2iadapter_openpose_sd14v1

4
0

t2iadapter_color_sd14v1

3
0

t2iadapter_sketch_sd14v1

3
0

t2iadapter_zoedepth_sd15v1

3
0

stable-diffusion-variants

2
2

t2iadapter_seg_sd14v1

2
0

open-sora-pipe-test

2
0

tiny-torch-full-checker

1
1

stable-diffusion-2-1-unclip-t2i-l

1
1

stable-diffusion-2-1-unclip-t2i-h

1
0

FLUX.1-Fill-dev-nf4

0
16

i2vgen-xl-diffusers

0
5

FLUX.1-Depth-dev-nf4

0
5

FLUX.1-Canny-dev-nf4

0
2

LTX-Video-0.9.1

0
2