camenduru

207 models • 1 total models in database
Sort by:

AnimateDiff

25,812
21

FLUX.1-dev-diffusers

14,551
4

damo-image-to-video

388
13

PASD

246
2

IDM-VTON-F16

235
11

Wonder3D

122
3

damo-video-to-video

108
4

plushies-pt

64
12

unianimate

51
8

FLUX.1-dev-ungated

`FLUX.1 [dev]` is a 12 billion parameter rectified flow transformer capable of generating images from text descriptions. For more information, please read our blog post. Key Features 1. Cutting-edge output quality, second only to our state-of-the-art model `FLUX.1 [pro]`. 2. Competitive prompt following, matching the performance of closed source alternatives . 3. Trained using guidance distillation, making `FLUX.1 [dev]` more efficient. 4. Open weights to drive new scientific research, and empower artists to develop innovative workflows. 5. Generated outputs can be used for personal, scientific, and commercial purposes as described in the [`FLUX.1 [dev]` Non-Commercial License](https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md). Usage We provide a reference implementation of `FLUX.1 [dev]`, as well as sampling code, in a dedicated github repository. Developers and creatives looking to build on top of `FLUX.1 [dev]` are encouraged to use this as a starting point. API Endpoints The FLUX.1 models are also available via API from the following sources - bfl.ml (currently `FLUX.1 [pro]`) - replicate.com - fal.ai - mystic.ai ComfyUI `FLUX.1 [dev]` is also available in Comfy UI for local inference with a node-based workflow. To use `FLUX.1 [dev]` with the 🧨 diffusers python library, first install or upgrade diffusers To learn more check out the diffusers documentation --- Limitations - This model is not intended or able to provide factual information. - As a statistical model this checkpoint might amplify existing societal biases. - The model may fail to generate output that matches the prompts. - Prompt following is heavily influenced by the prompting-style. Out-of-Scope Use The model and its derivatives may not be used - In any way that violates any applicable national, federal, state, local or international law or regulation. - For the purpose of exploiting, harming or attempting to exploit or harm minors in any way; including but not limited to the solicitation, creation, acquisition, or dissemination of child exploitative content. - To generate or disseminate verifiably false information and/or content with the purpose of harming others. - To generate or disseminate personal identifiable information that can be used to harm an individual. - To harass, abuse, threaten, stalk, or bully individuals or groups of individuals. - To create non-consensual nudity or illegal pornographic content. - For fully automated decision making that adversely impacts an individual's legal rights or otherwise creates or modifies a binding, enforceable obligation. - Generating or facilitating large-scale disinformation campaigns. License This model falls under the [`FLUX.1 [dev]` Non-Commercial License](https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md).

48
1

Video-to-Video

43
0

potat1

30
159

VistaDream

30
1

FLUX.1-Fill-dev-ungated

`FLUX.1 Fill [dev]` is a 12 billion parameter rectified flow transformer capable of filling areas in existing images based on a text description. For more information, please read our blog post. Key Features 1. Cutting-edge output quality, second only to our state-of-the-art model `FLUX.1 Fill [pro]`. 2. Blends impressive prompt following with completing the structure of your source image. 3. Trained using guidance distillation, making `FLUX.1 Fill [dev]` more efficient. 4. Open weights to drive new scientific research, and empower artists to develop innovative workflows. 5. Generated outputs can be used for personal, scientific, and commercial purposes as described in the [`FLUX.1 [dev]` Non-Commercial License](https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md). Usage We provide a reference implementation of `FLUX.1 Fill [dev]`, as well as sampling code, in a dedicated github repository. Developers and creatives looking to build on top of `FLUX.1 Fill [dev]` are encouraged to use this as a starting point. API Endpoints The FLUX.1 models are also available in our API bfl.ml To use `FLUX.1 Fill [dev]` with the 🧨 diffusers python library, first install or upgrade diffusers Then you can use `FluxFillPipeline` to run the model To learn more check out the diffusers documentation Limitations - This model is not intended or able to provide factual information. - As a statistical model this checkpoint might amplify existing societal biases. - The model may fail to generate output that matches the prompts. - Prompt following is heavily influenced by the prompting-style. - There may be slight-color shifts in areas that are not filled in - Filling in complex textures may produce lines at the edges of the filled-area. Out-of-Scope Use The model and its derivatives may not be used - In any way that violates any applicable national, federal, state, local or international law or regulation. - For the purpose of exploiting, harming or attempting to exploit or harm minors in any way; including but not limited to the solicitation, creation, acquisition, or dissemination of child exploitative content. - To generate or disseminate verifiably false information and/or content with the purpose of harming others. - To generate or disseminate personal identifiable information that can be used to harm an individual. - To harass, abuse, threat

27
0

VideoComposer

18
2

MeiGen-MultiTalk

MeiGen-MultiTalk • Audio-Driven Multi-Person Conversational Video Generation > We present MultiTalk, an open-source audio-driven multi-person conversational video generation model with the state-of-the-art lip synchronization accuracy. > ​​Key features:​​ > - 💬 ​​Realistic Conversations​​ - Supports single & multi-person generation > - 👥 ​​Interactive Character Control​​ - Direct virtual humans via prompts > - 🎤 ​​Generalization Performances​​ - Supports the generation of cartoon character and singing > - 📺 ​​Resolution Flexibility​​: 480p & 720p output at arbitrary aspect ratios > - ⏱️ Long Video Generation: Support video generation up to 15 seconds This repository hosts the model weights for MultiTalk. For installation, usage instructions, and further documentation, please visit our GitHub repository. Method We propose a novel framework, MultiTalk, for audio-driven multi-person conversational video generation. We investigate several schemes for audio injection and introduce the Label Rotary Position Embedding (L-RoPE) method. By assigning identical labels to audio embeddings and video latents, it effectively activates specific regions within the audio cross-attention map, thereby resolving incorrect binding issues. To localize the region of the specified person, we introduce the adaptive person localization by computing the similarity between the features of the given region of a person in the reference image and all the features of the whole video. Citation If you find our work helpful, please cite us. License Agreement The models in this repository are licensed under the Apache 2.0 License. We claim no rights over the your generated contents, granting you the freedom to use them while ensuring that your usage complies with the provisions of this license. You are fully accountable for your use of the models, which must not involve sharing any content that violates applicable laws, causes harm to individuals or groups, disseminates personal information intended for harm, spreads misinformation, or targets vulnerable populations.

license:apache-2.0
16
2

IMAGDressing

license:apache-2.0
16
0

xenmon-xl-v1

15
0

svd_xt_1_1_unet

12
0

DemoFusion

8
4

DreamClear

license:apache-2.0
8
1

Envision3D

7
0

InstantID

license:apache-2.0
6
0

TripoSR

5
1

plushies

4
22

MiniGPT4-7B

NaNK
llama
4
2

parakeet-rnnt-1.1b

NaNK
4
0

joy-caption-alpha-two

4
0

midstreet

3
0

ThemeStation

3
0

IICF

3
0

MiniGPT4

llama
2
5

champ

2
2

FLUX.1_Kontext-Lightning

Update 7/9/25: This model is now quantized and implemented in this example space. Seeing preliminary VRAM usage at around ~10GB with faster inferencing. Will be experimenting with different weights and schedulers to find particularly well-performing libraries. Highly experimental, will update with more details later. - 6-8 steps - Euler, SGM Uniform (Recommended, feel free to play around) Getting mixed results now, feel free to play around and share. Experimenting with FLUX.1-dev LoRAs and how it affects Kontext-dev. This model has been fused with acceleration LoRAs. License This model falls under the [FLUX.1 \[dev\] Non-Commercial License](https://github.com/black-forest-labs/flux/blob/main/modellicenses/LICENSE-FLUX1-dev), please familiarize yourself with the license.

2
1

xenmon-xl-v2

2
0

wonder3d-v1.0

license:agpl-3.0
2
0

marigold-e2e-ft-normals

license:apache-2.0
2
0

flux1-kontext-dev_fp8_e4m3fn_diffusers

2
0

XTTS-v1

1
4

test_gpu

1
1

EvoVLM-JP-v1-7B-4bit

NaNK
1
1

Meta-Llama-3.1-8B-Instruct

NaNK
llama
1
1

test-10

1
0

kosmos-2-patch14-224

1
0

IF-II-L-v1.0

1
0

Mixtral-8x22B-Instruct-v0.1

NaNK
1
0

RMBG-1.4

1
0

MythoMax-L2-13b

NaNK
llama
1
0

robustsam-vit-huge

license:mit
1
0

marigold-e2e-ft-depth

license:apache-2.0
1
0

SUPIR

0
165

FLUX.1-dev

0
142

Wav2Lip

0
68

YoloWorld-EfficientSAM

0
45

gaussian-splatting

0
40

SMPLer X

0
38

one-shot-talking-face

0
17

HandRefiner

0
11

stable-diffusion-3.5-large

0
9

xl_sliders

0
8

text2-video-zero

0
7

big-lama

0
7

dust3r

0
7

Diffutoon

0
7

instant-ngp

0
6

one-shot-talking-face-20.04-t4

0
6

one-shot-talking-face-20.04-a10

0
6

show

0
6

openpose

0
6

joy-caption-alpha-one

0
6

SadTalker

0
5

textdiffuser

0
5

video-retalking

0
5

OOTDiffusion

0
5

improved-aesthetic-predictor

0
4

potat1_dataset

0
4

3d-photo-inpainting

0
4

DWPose

0
4

ios-emoji-xl

0
4

beats

0
4

PeRF

0
4

dreamtalk

0
4

MagicDance

0
4

Multi-LoRA-Composition

0
4

PuLID

license:apache-2.0
0
4

one-shot-talking-face-20.04

0
3

DragGAN

0
3

StableSR

0
3

facechain

0
3

DeepFilterNet2

0
3

ProPainter

0
3

OpenVoice

0
3

cv_ddcolor_image-colorization

license:apache-2.0
0
3

Arc2Face

0
3

MuseTalk

0
3

CogVideoX-5b-8bit

NaNK
0
3

google_t5_v1.1

0
3

oasis-500m

license:mit
0
3

DimensionX

0
3

HunyuanVideo

0
3

memo

MEMO: Memory-Guided Diffusion for Expressive Talking Video Generation Longtao Zheng\, Yifan Zhang\, Hanzhong Guo\, Jiachun Pan, Zhenxiong Tan, Jiahao Lu, Chuanxin Tang, Bo An, Shuicheng Yan This repository contains the example inference script for the MEMO-preview model. The gif demo below is compressed. See our project page for full videos. > Our code will download the checkpoint from Hugging Face automatically, and the models for face analysis and vocal separation will be downloaded to `miscmodeldir` of `configs/inference.yaml`. If you want to download the models manually, please download the checkpoint from here and specify the path in `modelnameorpath` of `configs/inference.yaml`. > We tested the code on H100 and RTX 4090 GPUs using CUDA 12. Under the default settings (fps=30, inferencesteps=20), the inference time is around 1 second per frame on H100 and 2 seconds per frame on RTX 4090. We welcome community contributions to improve the inference speed or interfaces like ComfyUI. Our work is made possible thanks to high-quality open-source talking video datasets (including HDTF, VFHQ, CelebV-HQ, MultiTalk, and MEAD) and some pioneering works (such as EMO and Hallo). We acknowledge the potential of AI in generating talking videos, with applications spanning education, virtual assistants, and entertainment. However, we are equally aware of the ethical, legal, and societal challenges that misuse of this technology could pose. To reduce potential risks, we have only open-sourced a preview model for research purposes. Demos on our website use publicly available materials. We welcome copyright concerns—please contact us if needed, and we will address issues promptly. Users are required to ensure that their actions align with legal regulations, cultural norms, and ethical standards. It is strictly prohibited to use the model for creating malicious, misleading, defamatory, or privacy-infringing content, such as deepfake videos for political misinformation, impersonation, harassment, or fraud. We strongly encourage users to review generated content carefully, ensuring it meets ethical guidelines and respects the rights of all parties involved. Users must also ensure that their inputs (e.g., audio and reference images) and outputs are used with proper authorization. Unauthorized use of third-party intellectual property is strictly forbidden. While users may claim ownership of content generated by the model, they must ensure compliance with copyright laws, particularly when involving public figures' likeness, voice, or other aspects protected under personality rights. If you find our work useful, please use the following citation:

license:apache-2.0
0
3

openface-20.04-t4

0
2

text-to-video-synthesis

0
2

GroundingDINO

license:apache-2.0
0
2

roop

0
2

PanoHead

0
2

seamless-m4t-large

license:cc-by-nc-4.0
0
2

co-tracker

0
2

One-2-3-45

license:mit
0
2

DiffBIR

0
2

panic3d-anime-reconstruction

0
2

MiniGPT-5

0
2

sd-t2i-360panoimage

license:apache-2.0
0
2

Diff2Lip

0
2

AnimateAnyone

0
2

bucilianus-1

license:mit
0
2

sv3d

0
2

AdaSR-TalkingHead

0
2

StableVITON

0
2

ToonCrafter

0
2

Unique3D

0
2

EvTexture-2b

NaNK
0
2

LivePortrait

0
2

ultralytics

0
2

mdm

0
2

LivePortrait_InsightFace

0
2

PuppetMaster

0
2

stable-diffusion-3.5-large-turbo

0
2

FLUX.1-Krea-dev

0
2

Qwen-Loras

0
1

pocketsphinx-20.04-t4

0
1

tpu-train-tutorial-pt

0
1

tensor-rt

0
1

tensorrt-test

0
1

tensor-rt-sd14

0
1

ddpm

0
1

lora

0
1

xformers-0-1-7-t4

0
1

MeshDiffusion

0
1

instant-ngp-v2

0
1

SD-CN-Animation

0
1

hagrid-subsample-t1

0
1

hagrid-classification-512p

0
1

apex

0
1

NeMo

0
1

4D-Humans

0
1

howto100m

0
1

howto100m-json-test

0
1

Matting-Anything

0
1

pytorch3d-build

0
1

show-dataset

0
1

Rerender

0
1

MVDiffusion

0
1

shape_predictor_68_face_landmarks

0
1

StyleDrop

0
1

inst-inpaint

0
1

muavic

0
1

CoDeF

0
1

seamless-m4t-medium

license:cc-by-nc-4.0
0
1

CodeLlama-7b

NaNK
0
1

CodeLlama-13b

NaNK
0
1

vall-e-x

0
1

facechain-colab

0
1

GPEN

0
1

ffmpeg-cuda

0
1

SyncDreamer

0
1

Stable-Diffusion-NCNN

0
1

ncnn

0
1

NeuS2

0
1

4DGaussians

0
1

facexlib

0
1

GaussianDreamer

0
1

DiffSketcher

0
1

Wonder3D-test

0
1

SlimSAM

0
1

TalkingHead

0
1

UDiffText

0
1

AnyDoor

0
1

buddi

0
1

MoMask

0
1

anytext

0
1

3DTopia

0
1

image-sculpting

0
1

SketchVideo

0
1

pytorch_mgie

0
1

DSINE

0
1

trumans

0
1

EvoVLM-JP-v1-7B-8bit

NaNK
0
1

geowizard

0
1

BrushNet

0
1

IDM-VTON

0
1

Kandinsky3.1

0
1

MusePose

0
1

MeshAnything

0
1

chichi-pui

0
1

segment-anything-2

0
1

joy-caption

0
1

sapiens-body-part-segmentation

0
1

BiRefNet

0
1

clayify_test

0
1

GVHMR

0
1

CatVTON

license:cc-by-nc-sa-4.0
0
1

MimicMotion

0
1

Tora

0
1

EchoMimicV2

0
1

RAIN-v0.1

license:apache-2.0
0
1

wan2.1-14b-lora

NaNK
0
1