Trixster-MoE-14B-A8B-v1
2
llama
by
OccultAI
Language Model
OTHER
14B params
New
0 downloads
Early-stage
Edge AI:
Mobile
Laptop
Server
32GB+ RAM
Mobile
Laptop
Server
Quick Summary
AI model with specialized capabilities.
Device Compatibility
Mobile
4-6GB RAM
Laptop
16GB RAM
Server
GPU
Minimum Recommended
14GB+ RAM
Code Examples
System Prompt (Optional)yaml
# timeout /t 3 /nobreak && mergekit-yaml C:\mergekit-main\moe_karcher.yaml C:\mergekit-main\moe_karcher --copy-tokenizer --allow-crimes --out-shard-size 5B --trust-remote-code --lazy-unpickle --random-seed 420 --cuda
# MoE_Karcher example
# Blends corresponding experts across MoE models using geometric mean.
# MoE-aware Karcher merge that:
# 1. Identifies expert weights by pattern matching
# 2. Blends corresponding experts across MoE models
# 3. Handles router weights separately with optional strategies
merge_method: moe_karcher
base_model: B:\8B\Meme-Trix-MoE-14B-A8B-v1
models:
- model: B:\8B\Meme-Trix-MoE-14B-A8B-v1
- model: B:\8B\Babsie--CrossroadsLoki-MoE-2x8B
parameters:
max_iter: 1000
tol: 1e-9
router_strategy: karcher # Options: karcher, average, first, random_init
blend_experts: true # Blend corresponding experts (expert[0] + expert[0], etc.)
dtype: float32
out_dtype: bfloat16
tokenizer:
source: union
# chat_template: auto
name: Trixster-MoE-Karcher-14B-A8B-v1Deploy This Model
Production-ready deployment in minutes
Together.ai
Instant API access to this model
Production-ready inference API. Start free, scale to millions.
Try Free APIReplicate
One-click model deployment
Run models in the cloud with simple API. No DevOps required.
Deploy NowDisclosure: We may earn a commission from these partners. This helps keep LLMYourWay free.