LumiChats-Instruct-4B_lora
1
license:mit
by
adityakum667388
Language Model
OTHER
4B params
New
0 downloads
Early-stage
Edge AI:
Mobile
Laptop
Server
9GB+ RAM
Mobile
Laptop
Server
Quick Summary
AI model with specialized capabilities.
Device Compatibility
Mobile
4-6GB RAM
Laptop
16GB RAM
Server
GPU
Minimum Recommended
4GB+ RAM
Code Examples
š¶ Quick Startbash
pip install unsloth transformers bitsandbytes accelerate peftš¶ Quick Startpythontransformers
from unsloth import FastLanguageModel
from transformers import TextStreamer
model, tokenizer = FastLanguageModel.from_pretrained(
model_name = "adityakum667388/LumiChats-Instruct-14B_lora",
max_seq_length = 2048,
load_in_4bit = True,
)
FastLanguageModel.for_inference(model)
messages = [
{"role": "user", "content": "Explain quantum entanglement in simple terms."}
]
inputs = tokenizer.apply_chat_template(
messages,
tokenize = True,
add_generation_prompt = True,
return_tensors = "pt",
).to("cuda")
streamer = TextStreamer(tokenizer, skip_prompt=True)
_ = model.generate(
input_ids = inputs,
streamer = streamer,
max_new_tokens = 256,
temperature = 1.5,
min_p = 0.1,
use_cache = True,
)Deploy This Model
Production-ready deployment in minutes
Together.ai
Instant API access to this model
Production-ready inference API. Start free, scale to millions.
Try Free APIReplicate
One-click model deployment
Run models in the cloud with simple API. No DevOps required.
Deploy NowDisclosure: We may earn a commission from these partners. This helps keep LLMYourWay free.