Qwen2.5-1.5B-Instruct-kpi-tool-calling
17
1
license:apache-2.0
by
bhaiyahnsingh45
Other
OTHER
1.5B params
New
17 downloads
Early-stage
Edge AI:
Mobile
Laptop
Server
4GB+ RAM
Mobile
Laptop
Server
Quick Summary
AI model with specialized capabilities.
Device Compatibility
Mobile
4-6GB RAM
Laptop
16GB RAM
Server
GPU
Minimum Recommended
2GB+ RAM
Code Examples
Usagepythontransformers
from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel
import torch
import json
# Load model
base_model = AutoModelForCausalLM.from_pretrained(
"Qwen/Qwen2.5-1.5B-Instruct",
torch_dtype=torch.float16,
device_map="auto"
)
model = PeftModel.from_pretrained(base_model, "bhaiyahnsingh45/Qwen2.5-1.5B-Instruct-kpi-tool-calling")
tokenizer = AutoTokenizer.from_pretrained("bhaiyahnsingh45/Qwen2.5-1.5B-Instruct-kpi-tool-calling")
# Define tools schema
tools_json = '''
[
{
"type": "function",
"function": {
"name": "get_oee",
"description": "Get OEE (Overall Equipment Effectiveness) metrics",
"parameters": {
"type": "object",
"properties": {
"custom_start_date": {"type": "string", "description": "Start date (YYYY-MM-DD HH:MM:SS)"},
"custom_end_date": {"type": "string", "description": "End date (YYYY-MM-DD HH:MM:SS)"},
"machine": {"type": "string", "description": "Machine name"},
"line": {"type": "string", "description": "Production line"},
"plant": {"type": "string", "description": "Plant name"}
},
"required": ["custom_start_date", "custom_end_date"]
}
}
},
{
"type": "function",
"function": {
"name": "get_availability",
"description": "Get availability/uptime metrics",
"parameters": {
"type": "object",
"properties": {
"custom_start_date": {"type": "string", "description": "Start date (YYYY-MM-DD HH:MM:SS)"},
"custom_end_date": {"type": "string", "description": "End date (YYYY-MM-DD HH:MM:SS)"},
"machine": {"type": "string", "description": "Machine name"},
"line": {"type": "string", "description": "Production line"},
"plant": {"type": "string", "description": "Plant name"}
},
"required": ["custom_start_date", "custom_end_date"]
}
}
}
]
'''
tools = json.loads(tools_json)
# System prompt
system_prompt = "You are a function calling assistant for manufacturing KPI data. Respond ONLY with function calls."
# Example query
user_query = "Show me the OEE for LINE_A01 from January 1st to January 31st 2024"
# Format messages
messages = [
{"role": "system", "content": system_prompt},
{"role": "user", "content": user_query}
]
# Generate response
text = tokenizer.apply_chat_template(messages, tools=tools, tokenize=False, add_generation_prompt=True)
inputs = tokenizer([text], return_tensors="pt").to(model.device)
with torch.no_grad():
outputs = model.generate(**inputs, max_new_tokens=512, temperature=0.1, do_sample=True)
response = tokenizer.decode(outputs[0][inputs.input_ids.shape[1]:], skip_special_tokens=True)
print(response)Deploy This Model
Production-ready deployment in minutes
Together.ai
Instant API access to this model
Production-ready inference API. Start free, scale to millions.
Try Free APIReplicate
One-click model deployment
Run models in the cloud with simple API. No DevOps required.
Deploy NowDisclosure: We may earn a commission from these partners. This helps keep LLMYourWay free.