Trixster-MoE-14B-A8B-v1

2
llama
by
OccultAI
Language Model
OTHER
14B params
New
0 downloads
Early-stage
Edge AI:
Mobile
Laptop
Server
32GB+ RAM
Mobile
Laptop
Server
Quick Summary

AI model with specialized capabilities.

Device Compatibility

Mobile
4-6GB RAM
Laptop
16GB RAM
Server
GPU
Minimum Recommended
14GB+ RAM

Code Examples

System Prompt (Optional)yaml
# timeout /t 3 /nobreak && mergekit-yaml C:\mergekit-main\moe_karcher.yaml C:\mergekit-main\moe_karcher --copy-tokenizer --allow-crimes --out-shard-size 5B --trust-remote-code --lazy-unpickle --random-seed 420 --cuda

# MoE_Karcher example
# Blends corresponding experts across MoE models using geometric mean.
# MoE-aware Karcher merge that:
#    1. Identifies expert weights by pattern matching
#    2. Blends corresponding experts across MoE models
#    3. Handles router weights separately with optional strategies
merge_method: moe_karcher
base_model: B:\8B\Meme-Trix-MoE-14B-A8B-v1
models:
  - model: B:\8B\Meme-Trix-MoE-14B-A8B-v1
  - model: B:\8B\Babsie--CrossroadsLoki-MoE-2x8B
parameters:
  max_iter: 1000
  tol: 1e-9
  router_strategy: karcher  # Options: karcher, average, first, random_init
  blend_experts: true  # Blend corresponding experts (expert[0] + expert[0], etc.)
dtype: float32
out_dtype: bfloat16
tokenizer:
  source: union
# chat_template: auto
name: Trixster-MoE-Karcher-14B-A8B-v1

Deploy This Model

Production-ready deployment in minutes

Together.ai

Instant API access to this model

Fastest API

Production-ready inference API. Start free, scale to millions.

Try Free API

Replicate

One-click model deployment

Easiest Setup

Run models in the cloud with simple API. No DevOps required.

Deploy Now

Disclosure: We may earn a commission from these partners. This helps keep LLMYourWay free.