MIRA-Large
1
license:mit
by
modeluser8888
Other
OTHER
New
1 downloads
Early-stage
Edge AI:
Mobile
Laptop
Server
Unknown
Mobile
Laptop
Server
Quick Summary
AI model with specialized capabilities.
Code Examples
Usagepythonpytorch
import torch
from MIRA.mira.models.modeling_mira import MIRAForPrediction
from MIRA.mira.models.utils_time_normalization import normalize_time_for_ctrope
# Load the pre-trained model
model = MIRAForPrediction.from_pretrained(ckpt_path).cuda()
model.eval()
# Example inference (pseudo-code)
device = next(model.parameters()).device
hist_vals = hist_vals.to(device)
hist_times = hist_times.to(device)
future_times = future_times.to(device)
cur_vals = hist_vals.clone()
cur_times = hist_times.clone()
preds_norm = []
for i in range(P):
# model input
inp_vals = cur_vals.unsqueeze(-1) # [1, L, 1]
inp_times = cur_times # [1, L]
with torch.no_grad():
out = model(
input_ids=inp_vals,
time_values=inp_times,
next_target_time_values=None, # no ODE for 1-step
return_dict=True,
)
next_norm = out.logits[:, -1, :] # [1, 1]
preds_norm.append(next_norm.squeeze(0))
next_t = future_times[:, i:i+1]
cur_vals = torch.cat([cur_vals, next_norm], dim=1)
cur_times = torch.cat([cur_times, next_t], dim=1)
preds_norm = torch.stack(preds_norm, dim=1) # [1, P]
preds = preds_norm * std[:, :, :] + mean[:, :, :]
preds = preds.squeeze(0)
print(preds)Deploy This Model
Production-ready deployment in minutes
Together.ai
Instant API access to this model
Production-ready inference API. Start free, scale to millions.
Try Free APIReplicate
One-click model deployment
Run models in the cloud with simple API. No DevOps required.
Deploy NowDisclosure: We may earn a commission from these partners. This helps keep LLMYourWay free.