yujiepan
stable-diffusion-3-tiny-random
tiny-random-swin-patch4-window7-224
clip-vit-tiny-random-patch14-336
qwen3-tiny-random-tp
opt-tiny-random
qwen3-tiny-random
qwen2.5-tiny-random
This model is for debugging. It is randomly initialized using the config from Qwen/Qwen2.5-72B-Instruct but with smaller size.
gemma-3-tiny-random
gemma-tiny-random
smollm-tiny-random
qwen2-tiny-random
mistral-v0.3-tiny-random
ring-tiny-random
This tiny model is intended for debugging. It is randomly initialized using the configuration adapted from inclusionAI/Ring-1T-preview.
glm-5-tiny-random
qwen2.5-128k-tiny-random
phi-4-tiny-random
meta-llama-3-tiny-random
llama-2-tiny-random
qwen1.5-tiny-random
deepseek-v3-tiny-random
llama-2-tiny-3layers-random
llama-3-tiny-random
gemma-2-tiny-random
deepseek-llm-tiny-random
meta-llama-3.1-tiny-random
qwen3-next-moe-tiny-random
llama-3.2-tiny-random
mixtral-8xtiny-random
QwQ-preview-tiny-random
bloom-tiny-random
glm-4-tiny-random
qwen1.5-moe-tiny-random
glm-moe-dsa-tiny-random
gptj-tiny-random
mixtral-tiny-random
qwen3-moe-tiny-random
llama-3.3-tiny-random
QwQ-tiny-random
dbrx-tiny-random
llama-3.1-tiny-random
dbrx-tiny256-random
starcoder-tiny-random
opt-tiny-2layers-random
mistral-tiny-random
mistral-nemo-2407-tiny-random
meta-llama-3.1-tiny-random-hidden128
meta-llama-3.2-tiny-random
stablelm-2-tiny-random
mathstral-v0.1-tiny-random
deepseek-v2-0628-tiny-random
qwq-tiny-random-dim64
gpt-oss-tiny-random
This tiny model is for debugging. It is randomly initialized with the config adapted from openai/gpt-oss-120b. Note: This model is in BF16; quantized MXFP4 FFN is not used.
llama-3.3-tiny-random-dim64
This tiny model is for debugging. It is randomly initialized with the config adapted from meta-llama/Llama-3.3-70B-Instruct.
glm-4.5-tiny-random
This tiny model is for debugging. It is randomly initialized with the config adapted from zai-org/GLM-4.5. Note: The `transformers` implementation does not have multi-token prediction (MTP) support. So you might see some "weights not loaded" warnings. This is expected.
gpt-oss-tiny-random-mxfp4
glm-4-moe-tiny-random
This tiny model is for debugging. It is randomly initialized with the config adapted from zai-org/GLM-4.5. Note: The `transformers` implementation does not have multi-token prediction (MTP) support. So you might see some "weights not loaded" warnings. This is expected.
gpt-oss-tiny-random-bf16
This tiny model is for debugging. It is randomly initialized with the config adapted from openai/gpt-oss-120b. Note: This model is in BF16; quantized MXFP4 FFN is not used.
smollm3-tiny-random
This tiny model is for debugging. It is randomly initialized with the config adapted from HuggingFaceTB/SmolLM3-3B.
ernie-4.5-moe-tiny-random
This tiny model is intended for debugging. It is randomly initialized using the configuration adapted from baidu/ERNIE-4.5-21B-A3B-Thinking.
seed-oss-tiny-random
llama-3.2-vision-tiny-random
llama-4-tiny-random
deepseek-v2-tiny-random
falcon-tiny-random
qwen2-vl-tiny-random
falcon-new-tiny64-random
mamba2-tiny-random
phi-3-tiny-random
jamba-tiny-random
phi-3.5-tiny-random
mamba2-codestral-v0.1-tiny-random
llama-4-8E-tiny-random
codestral-v0.1-tiny-random
mamba-tiny-random
kimi-k2-tiny-random
This tiny model is for debugging. It is randomly initialized with the config adapted from moonshotai/Kimi-K2-Instruct.
phi-moe-tiny-random
phi-3.5-moe-tiny-random
gemma-4-e-tiny-random
falcon-new-tiny-random
whisper-v3-tiny-random
This model is for debugging. It is randomly initialized with the config from openai/whisper-large-v3 but is of smaller size.
tiny-random-bert
hunyuan-dense-v1-tiny-random
This tiny model is for debugging. It is randomly initialized with the config adapted from tencent/Hunyuan-7B-Instruct.
qwen2.5-omni-tiny-random
bamba-tiny-random
hunyuan-tiny-random
This tiny model is for debugging. It is randomly initialized with the config adapted from tencent/Hunyuan-7B-Instruct.
grok-1-tiny-random
internlm2-tiny-random
deepseek-v3.1-tiny-random
This tiny model is for debugging. It is randomly initialized with the config adapted from deepseek-ai/DeepSeek-V3.1.
apertus-tiny-random
This tiny model is intended for debugging. It is randomly initialized using the configuration adapted from swiss-ai/Apertus-70B-Instruct-2509.
falcon-new-tiny-random-awq-w4g64
gemma-4e-tiny-random
falcon-mamba-tiny-random
hymba-tiny-random
chatglm3-tiny-random
llama-3-tiny-random-gptq-w4
mpt-tiny-random
qwen-vl-tiny-random
phi-3-vision-tiny-random
qwen2-audio-tiny-random
qvq-preview-tiny-random
step3-tiny-random-vllm
This tiny model is for debugging. It is randomly initialized with the config adapted from stepfun-ai/step3. Note: if you want the model version that follows transformers' naming, see the model without "-vllm" suffix.
tiny-random-SwinModel
jamba-1.5-tiny-random
minimax-m1-tiny-random
gemma-3n-tiny-random-dim4
hunyuan-moe-tiny-random
This tiny model is for debugging. It is randomly initialized with the config adapted from tencent/Hunyuan-A13B-Instruct.
gemma-3n-tiny-random
minicpm4-tiny-random
This tiny model is for debugging. It is randomly initialized with the config adapted from openbmb/MiniCPM4-8B.
ernie-4.5-tiny-random
This tiny model is intended for debugging. It is randomly initialized using the configuration adapted from baidu/ERNIE-4.5-0.3B-PT.
lfm2-tiny-random
This tiny model is for debugging. It is randomly initialized with the config adapted from LiquidAI/LFM2-1.2B.
phi-4-multimodal-tiny-random
mixtral-8xtiny-random-openvino-8bit
gemma-4-dense-tiny-random
meta-llama-3.1-tiny-random-hidden128-awq-w4g64
gemma-4-moe-tiny-random
minicpm-v-4-tiny-random
glm-4.1v-tiny-random
longcat-flash-tiny-random
This tiny model is for debugging. It is randomly initialized with the config adapted from meituan-longcat/LongCat-Flash-Chat.
glm-4v-tiny-random
voxtral-tiny-random
phi-4-flash-tiny-random
This tiny model is for debugging. It is randomly initialized with the config adapted from microsoft/Phi-4-mini-flash-reasoning.
qwen3-vl-moe-tiny-random
This tiny model is intended for debugging. It is randomly initialized using the configuration adapted from Qwen/Qwen3-VL-235B-A22B-Instruct.
step3-tiny-random
This tiny model is for debugging. It is randomly initialized with the config adapted from stepfun-ai/step3. Note: For vLLM supported version, see yujiepan/step3-tiny-random-vllm.
dreamshaper-8-lcm-openvino-w8a8
sam3-tiny-random
minicpm4.1-tiny-random
This tiny model is intended for debugging. It is randomly initialized using the configuration adapted from openbmb/MiniCPM4.1-8B.
glm-4.5v-tiny-random
minimax-m2-tiny-random
glm-4v-moe-tiny-random
llava-onevision-1.5-tiny-random
bailing-moe-v2-tiny-random
This tiny model is intended for debugging. It is randomly initialized using the configuration adapted from inclusionAI/Ring-1T-preview.
apriel-1.5-tiny-random
kimi-k2.5-tiny-random
minicpm-v-4_5-tiny-random
ui-tars-1.5-7B-GPTQ-W4A16g128
granite-moe-hybrid-tiny-random
This tiny model is intended for debugging. It is randomly initialized using the configuration adapted from ibm-granite/granite-4.0-h-small.
qwen3-vl-tiny-random
kormo-tiny-random
ernie-4.5-vl-moe-tiny-random
granite-4.0-h-tiny-random
This tiny model is intended for debugging. It is randomly initialized using the configuration adapted from ibm-granite/granite-4.0-h-small.
ui-tars-1.5-7B-bf16
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: [More Information Needed] - Funded by [optional]: [More Information Needed] - Shared by [optional]: [More Information Needed] - Model type: [More Information Needed] - Language(s) (NLP): [More Information Needed] - License: [More Information Needed] - Finetuned from model [optional]: [More Information Needed] - Repository: [More Information Needed] - Paper [optional]: [More Information Needed] - Demo [optional]: [More Information Needed] Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: [More Information Needed] - Hours used: [More Information Needed] - Cloud Provider: [More Information Needed] - Compute Region: [More Information Needed] - Carbon Emitted: [More Information Needed]