YanoljaNEXT-Rosetta-4B-2511-GGUF
150
4.0B
32 languages
BF16
llama-cpp
by
yanolja
Language Model
OTHER
4B params
New
150 downloads
Early-stage
Edge AI:
Mobile
Laptop
Server
9GB+ RAM
Mobile
Laptop
Server
Quick Summary
AI model with specialized capabilities.
Device Compatibility
Mobile
4-6GB RAM
Laptop
16GB RAM
Server
GPU
Minimum Recommended
4GB+ RAM
Code Examples
How to usebashllama.cpp
# Example: use a Q5_K_M quantized file (adjust the path/model to your choice)
MODEL="path/to/YanoljaNEXT-Rosetta-4B-2511-q5_k_m.gguf"
# Build a formatted prompt using the included chat template roles
# (see release/YanoljaNEXT-Rosetta-4B-2511/chat_template.jinja)
read -r -d '' PROMPT <<'EOT'
<start_of_turn>instruction
Translate the user's text to Korean. Keep the JSON structure and keys.
Context: Simple introduction about a tech company.
Tone: Informative and helpful
Glossary:
- Yanolja NEXT -> 야놀자넥스트
- travel industry -> 여행 산업
Output format: JSON
Provide the final translation immediately without any other text.
<end_of_turn>
<start_of_turn>source
{"company_name": "Yanolja NEXT", "description": "Yanolja NEXT is a company that provides cutting-edge technology for the global travel industry."}
<end_of_turn>
<start_of_turn>translation\n
EOT
# Run llama.cpp (adjust -n/-c/--temp as needed)
llama-cli -m "$MODEL" -p "$PROMPT" -n 64 -c 4096 --temp 0.7 -no-cnvRun llama.cpp (adjust -n/-c/--temp as needed)bash
MODEL="path/to/YanoljaNEXT-Rosetta-4B-2511-q5_k_m.gguf"
llama-server -m "$MODEL" -c 4096 --host 0.0.0.0 --port 8080Deploy This Model
Production-ready deployment in minutes
Together.ai
Instant API access to this model
Production-ready inference API. Start free, scale to millions.
Try Free APIReplicate
One-click model deployment
Run models in the cloud with simple API. No DevOps required.
Deploy NowDisclosure: We may earn a commission from these partners. This helps keep LLMYourWay free.