Qwen3-0.6B-Sushi-Math-Code-Expert
45
1
—
by
gss1147
Language Model
OTHER
0.6B params
New
45 downloads
Early-stage
Edge AI:
Mobile
Laptop
Server
2GB+ RAM
Mobile
Laptop
Server
Quick Summary
AI model with specialized capabilities.
Device Compatibility
Mobile
4-6GB RAM
Laptop
16GB RAM
Server
GPU
Minimum Recommended
1GB+ RAM
Code Examples
Folder Structuretext
qwen3-sushi-math-code-expert/
├── main.py # Core Python script for model loading, inference pipeline, and query handling
├── requirements.txt # Dependencies for correct implementation
├── config.yaml # Configuration for model, device, and pipeline settings
├── prompts.json # JSON file for predefined prompt templates (e.g., thinking mode)
├── logs/ # Folder for runtime logs (created dynamically)
│ └── inference.log # TXT log file (appended during runtime)
└── db/ # Folder for simple SQLite DB for query history
└── history.db # SQLite DB file (created dynamically)SQLite DB file (created dynamically)yaml
model:
name: "gss1147/Qwen3-0.6B-Sushi-Math-Code-Expert"
dtype: "float16"
trust_remote_code: true
pipeline:
max_length: 512
temperature: 0.7
top_p: 0.9
thinking_mode: true # Enable thinking mode for math/code reasoning
device:
type: "cuda" # Use "cpu" if no GPU
logging:
log_file: "logs/inference.log"
db_file: "db/history.db"Enable thinking mode for math/code reasoningjson
{
"thinking_mode": "You are a math and code expert. Use /think to enable thinking mode for complex reasoning. Query: {query}",
"non_thinking_mode": "You are a general assistant. Use /no_think for efficient response. Query: {query}"
}Deploy This Model
Production-ready deployment in minutes
Together.ai
Instant API access to this model
Production-ready inference API. Start free, scale to millions.
Try Free APIReplicate
One-click model deployment
Run models in the cloud with simple API. No DevOps required.
Deploy NowDisclosure: We may earn a commission from these partners. This helps keep LLMYourWay free.