First_Persian_SLM_Big_Update_Version3_ysnrfd
7
2
1 language
—
by
ysn-rfd
Language Model
OTHER
New
7 downloads
Early-stage
Edge AI:
Mobile
Laptop
Server
Unknown
Mobile
Laptop
Server
Quick Summary
WARNINNGS: This Model IS Pre-Trained, in the future will be finetuned.
Code Examples
Model Architecture and Objectivepythontransformers
import torch
from transformers import GPT2LMHeadModel, GPT2Tokenizer
import re
model_path = "./Path_To_Model"
print(f"{model_path}...")
tokenizer = GPT2Tokenizer.from_pretrained(model_path)
model = GPT2LMHeadModel.from_pretrained(model_path)
device = "cuda" if torch.cuda.is_available() else "cpu"
print(f"{device}")
model.to(device)
model.eval()
try:
end_token_id = tokenizer.convert_tokens_to_ids("### End")
if end_token_id == tokenizer.unk_token_id:
end_token_id = None
except:
end_token_id = None
print("\n------+------+------+------ model is ready for testing +------+------++------\n")
print("type exit for exit")
print("\nYSNRFD")
print("------+------+------+----------------------+------+------++------\n")
while True:
user_input = input("\nYou: ").strip()
if user_input.lower() in ["exit"]:
print("\n good bye (ysnrfd)")
break
prompt = f"### Human: {user_input}\n### Assistant:"
inputs = tokenizer(
prompt,
return_tensors="pt",
truncation=True,
max_length=512
).to(device)
with torch.no_grad():
outputs = model.generate(
**inputs,
max_new_tokens=400,
temperature=0.7,
top_p=0.9,
do_sample=True,
pad_token_id=tokenizer.eos_token_id,
eos_token_id=end_token_id or tokenizer.eos_token_id,
repetition_penalty=1.05
)
full_response = tokenizer.decode(outputs[0], skip_special_tokens=False)
assistant_response = full_response[len(prompt):].strip()
if "### End" in assistant_response:
assistant_response = assistant_response.split("### End")[0].strip()
assistant_response = re.sub(r'^###\s*Assistant:\s*', '', assistant_response)
if assistant_response:
print(f"\nBot: {assistant_response}")
else:
print("\nBot: please again say")Model Architecture and Objectivepythontransformers
import torch
from transformers import GPT2LMHeadModel, GPT2Tokenizer
import re
model_path = "./Path_To_Model"
print(f"{model_path}...")
tokenizer = GPT2Tokenizer.from_pretrained(model_path)
model = GPT2LMHeadModel.from_pretrained(model_path)
device = "cuda" if torch.cuda.is_available() else "cpu"
print(f"{device}")
model.to(device)
model.eval()
try:
end_token_id = tokenizer.convert_tokens_to_ids("### End")
if end_token_id == tokenizer.unk_token_id:
end_token_id = None
except:
end_token_id = None
print("\n------+------+------+------ model is ready for testing +------+------++------\n")
print("type exit for exit")
print("\nYSNRFD")
print("------+------+------+----------------------+------+------++------\n")
while True:
user_input = input("\nYou: ").strip()
if user_input.lower() in ["exit"]:
print("\n good bye (ysnrfd)")
break
prompt = f"### Human: {user_input}\n### Assistant:"
inputs = tokenizer(
prompt,
return_tensors="pt",
truncation=True,
max_length=512
).to(device)
with torch.no_grad():
outputs = model.generate(
**inputs,
max_new_tokens=400,
temperature=0.7,
top_p=0.9,
do_sample=True,
pad_token_id=tokenizer.eos_token_id,
eos_token_id=end_token_id or tokenizer.eos_token_id,
repetition_penalty=1.05
)
full_response = tokenizer.decode(outputs[0], skip_special_tokens=False)
assistant_response = full_response[len(prompt):].strip()
if "### End" in assistant_response:
assistant_response = assistant_response.split("### End")[0].strip()
assistant_response = re.sub(r'^###\s*Assistant:\s*', '', assistant_response)
if assistant_response:
print(f"\nBot: {assistant_response}")
else:
print("\nBot: please again say")Model Architecture and Objectivepythontransformers
import torch
from transformers import GPT2LMHeadModel, GPT2Tokenizer
import re
model_path = "./Path_To_Model"
print(f"{model_path}...")
tokenizer = GPT2Tokenizer.from_pretrained(model_path)
model = GPT2LMHeadModel.from_pretrained(model_path)
device = "cuda" if torch.cuda.is_available() else "cpu"
print(f"{device}")
model.to(device)
model.eval()
try:
end_token_id = tokenizer.convert_tokens_to_ids("### End")
if end_token_id == tokenizer.unk_token_id:
end_token_id = None
except:
end_token_id = None
print("\n------+------+------+------ model is ready for testing +------+------++------\n")
print("type exit for exit")
print("\nYSNRFD")
print("------+------+------+----------------------+------+------++------\n")
while True:
user_input = input("\nYou: ").strip()
if user_input.lower() in ["exit"]:
print("\n good bye (ysnrfd)")
break
prompt = f"### Human: {user_input}\n### Assistant:"
inputs = tokenizer(
prompt,
return_tensors="pt",
truncation=True,
max_length=512
).to(device)
with torch.no_grad():
outputs = model.generate(
**inputs,
max_new_tokens=400,
temperature=0.7,
top_p=0.9,
do_sample=True,
pad_token_id=tokenizer.eos_token_id,
eos_token_id=end_token_id or tokenizer.eos_token_id,
repetition_penalty=1.05
)
full_response = tokenizer.decode(outputs[0], skip_special_tokens=False)
assistant_response = full_response[len(prompt):].strip()
if "### End" in assistant_response:
assistant_response = assistant_response.split("### End")[0].strip()
assistant_response = re.sub(r'^###\s*Assistant:\s*', '', assistant_response)
if assistant_response:
print(f"\nBot: {assistant_response}")
else:
print("\nBot: please again say")Model Architecture and Objectivepythontransformers
import torch
from transformers import GPT2LMHeadModel, GPT2Tokenizer
import re
model_path = "./Path_To_Model"
print(f"{model_path}...")
tokenizer = GPT2Tokenizer.from_pretrained(model_path)
model = GPT2LMHeadModel.from_pretrained(model_path)
device = "cuda" if torch.cuda.is_available() else "cpu"
print(f"{device}")
model.to(device)
model.eval()
try:
end_token_id = tokenizer.convert_tokens_to_ids("### End")
if end_token_id == tokenizer.unk_token_id:
end_token_id = None
except:
end_token_id = None
print("\n------+------+------+------ model is ready for testing +------+------++------\n")
print("type exit for exit")
print("\nYSNRFD")
print("------+------+------+----------------------+------+------++------\n")
while True:
user_input = input("\nYou: ").strip()
if user_input.lower() in ["exit"]:
print("\n good bye (ysnrfd)")
break
prompt = f"### Human: {user_input}\n### Assistant:"
inputs = tokenizer(
prompt,
return_tensors="pt",
truncation=True,
max_length=512
).to(device)
with torch.no_grad():
outputs = model.generate(
**inputs,
max_new_tokens=400,
temperature=0.7,
top_p=0.9,
do_sample=True,
pad_token_id=tokenizer.eos_token_id,
eos_token_id=end_token_id or tokenizer.eos_token_id,
repetition_penalty=1.05
)
full_response = tokenizer.decode(outputs[0], skip_special_tokens=False)
assistant_response = full_response[len(prompt):].strip()
if "### End" in assistant_response:
assistant_response = assistant_response.split("### End")[0].strip()
assistant_response = re.sub(r'^###\s*Assistant:\s*', '', assistant_response)
if assistant_response:
print(f"\nBot: {assistant_response}")
else:
print("\nBot: please again say")Model Architecture and Objectivepythontransformers
import torch
from transformers import GPT2LMHeadModel, GPT2Tokenizer
import re
model_path = "./Path_To_Model"
print(f"{model_path}...")
tokenizer = GPT2Tokenizer.from_pretrained(model_path)
model = GPT2LMHeadModel.from_pretrained(model_path)
device = "cuda" if torch.cuda.is_available() else "cpu"
print(f"{device}")
model.to(device)
model.eval()
try:
end_token_id = tokenizer.convert_tokens_to_ids("### End")
if end_token_id == tokenizer.unk_token_id:
end_token_id = None
except:
end_token_id = None
print("\n------+------+------+------ model is ready for testing +------+------++------\n")
print("type exit for exit")
print("\nYSNRFD")
print("------+------+------+----------------------+------+------++------\n")
while True:
user_input = input("\nYou: ").strip()
if user_input.lower() in ["exit"]:
print("\n good bye (ysnrfd)")
break
prompt = f"### Human: {user_input}\n### Assistant:"
inputs = tokenizer(
prompt,
return_tensors="pt",
truncation=True,
max_length=512
).to(device)
with torch.no_grad():
outputs = model.generate(
**inputs,
max_new_tokens=400,
temperature=0.7,
top_p=0.9,
do_sample=True,
pad_token_id=tokenizer.eos_token_id,
eos_token_id=end_token_id or tokenizer.eos_token_id,
repetition_penalty=1.05
)
full_response = tokenizer.decode(outputs[0], skip_special_tokens=False)
assistant_response = full_response[len(prompt):].strip()
if "### End" in assistant_response:
assistant_response = assistant_response.split("### End")[0].strip()
assistant_response = re.sub(r'^###\s*Assistant:\s*', '', assistant_response)
if assistant_response:
print(f"\nBot: {assistant_response}")
else:
print("\nBot: please again say")Model Architecture and Objectivepythontransformers
import torch
from transformers import GPT2LMHeadModel, GPT2Tokenizer
import re
model_path = "./Path_To_Model"
print(f"{model_path}...")
tokenizer = GPT2Tokenizer.from_pretrained(model_path)
model = GPT2LMHeadModel.from_pretrained(model_path)
device = "cuda" if torch.cuda.is_available() else "cpu"
print(f"{device}")
model.to(device)
model.eval()
try:
end_token_id = tokenizer.convert_tokens_to_ids("### End")
if end_token_id == tokenizer.unk_token_id:
end_token_id = None
except:
end_token_id = None
print("\n------+------+------+------ model is ready for testing +------+------++------\n")
print("type exit for exit")
print("\nYSNRFD")
print("------+------+------+----------------------+------+------++------\n")
while True:
user_input = input("\nYou: ").strip()
if user_input.lower() in ["exit"]:
print("\n good bye (ysnrfd)")
break
prompt = f"### Human: {user_input}\n### Assistant:"
inputs = tokenizer(
prompt,
return_tensors="pt",
truncation=True,
max_length=512
).to(device)
with torch.no_grad():
outputs = model.generate(
**inputs,
max_new_tokens=400,
temperature=0.7,
top_p=0.9,
do_sample=True,
pad_token_id=tokenizer.eos_token_id,
eos_token_id=end_token_id or tokenizer.eos_token_id,
repetition_penalty=1.05
)
full_response = tokenizer.decode(outputs[0], skip_special_tokens=False)
assistant_response = full_response[len(prompt):].strip()
if "### End" in assistant_response:
assistant_response = assistant_response.split("### End")[0].strip()
assistant_response = re.sub(r'^###\s*Assistant:\s*', '', assistant_response)
if assistant_response:
print(f"\nBot: {assistant_response}")
else:
print("\nBot: please again say")Model Architecture and Objectivepythontransformers
import torch
from transformers import GPT2LMHeadModel, GPT2Tokenizer
import re
model_path = "./Path_To_Model"
print(f"{model_path}...")
tokenizer = GPT2Tokenizer.from_pretrained(model_path)
model = GPT2LMHeadModel.from_pretrained(model_path)
device = "cuda" if torch.cuda.is_available() else "cpu"
print(f"{device}")
model.to(device)
model.eval()
try:
end_token_id = tokenizer.convert_tokens_to_ids("### End")
if end_token_id == tokenizer.unk_token_id:
end_token_id = None
except:
end_token_id = None
print("\n------+------+------+------ model is ready for testing +------+------++------\n")
print("type exit for exit")
print("\nYSNRFD")
print("------+------+------+----------------------+------+------++------\n")
while True:
user_input = input("\nYou: ").strip()
if user_input.lower() in ["exit"]:
print("\n good bye (ysnrfd)")
break
prompt = f"### Human: {user_input}\n### Assistant:"
inputs = tokenizer(
prompt,
return_tensors="pt",
truncation=True,
max_length=512
).to(device)
with torch.no_grad():
outputs = model.generate(
**inputs,
max_new_tokens=400,
temperature=0.7,
top_p=0.9,
do_sample=True,
pad_token_id=tokenizer.eos_token_id,
eos_token_id=end_token_id or tokenizer.eos_token_id,
repetition_penalty=1.05
)
full_response = tokenizer.decode(outputs[0], skip_special_tokens=False)
assistant_response = full_response[len(prompt):].strip()
if "### End" in assistant_response:
assistant_response = assistant_response.split("### End")[0].strip()
assistant_response = re.sub(r'^###\s*Assistant:\s*', '', assistant_response)
if assistant_response:
print(f"\nBot: {assistant_response}")
else:
print("\nBot: please again say")Model Architecture and Objectivepythontransformers
import torch
from transformers import GPT2LMHeadModel, GPT2Tokenizer
import re
model_path = "./Path_To_Model"
print(f"{model_path}...")
tokenizer = GPT2Tokenizer.from_pretrained(model_path)
model = GPT2LMHeadModel.from_pretrained(model_path)
device = "cuda" if torch.cuda.is_available() else "cpu"
print(f"{device}")
model.to(device)
model.eval()
try:
end_token_id = tokenizer.convert_tokens_to_ids("### End")
if end_token_id == tokenizer.unk_token_id:
end_token_id = None
except:
end_token_id = None
print("\n------+------+------+------ model is ready for testing +------+------++------\n")
print("type exit for exit")
print("\nYSNRFD")
print("------+------+------+----------------------+------+------++------\n")
while True:
user_input = input("\nYou: ").strip()
if user_input.lower() in ["exit"]:
print("\n good bye (ysnrfd)")
break
prompt = f"### Human: {user_input}\n### Assistant:"
inputs = tokenizer(
prompt,
return_tensors="pt",
truncation=True,
max_length=512
).to(device)
with torch.no_grad():
outputs = model.generate(
**inputs,
max_new_tokens=400,
temperature=0.7,
top_p=0.9,
do_sample=True,
pad_token_id=tokenizer.eos_token_id,
eos_token_id=end_token_id or tokenizer.eos_token_id,
repetition_penalty=1.05
)
full_response = tokenizer.decode(outputs[0], skip_special_tokens=False)
assistant_response = full_response[len(prompt):].strip()
if "### End" in assistant_response:
assistant_response = assistant_response.split("### End")[0].strip()
assistant_response = re.sub(r'^###\s*Assistant:\s*', '', assistant_response)
if assistant_response:
print(f"\nBot: {assistant_response}")
else:
print("\nBot: please again say")Model Architecture and Objectivepythontransformers
import torch
from transformers import GPT2LMHeadModel, GPT2Tokenizer
import re
model_path = "./Path_To_Model"
print(f"{model_path}...")
tokenizer = GPT2Tokenizer.from_pretrained(model_path)
model = GPT2LMHeadModel.from_pretrained(model_path)
device = "cuda" if torch.cuda.is_available() else "cpu"
print(f"{device}")
model.to(device)
model.eval()
try:
end_token_id = tokenizer.convert_tokens_to_ids("### End")
if end_token_id == tokenizer.unk_token_id:
end_token_id = None
except:
end_token_id = None
print("\n------+------+------+------ model is ready for testing +------+------++------\n")
print("type exit for exit")
print("\nYSNRFD")
print("------+------+------+----------------------+------+------++------\n")
while True:
user_input = input("\nYou: ").strip()
if user_input.lower() in ["exit"]:
print("\n good bye (ysnrfd)")
break
prompt = f"### Human: {user_input}\n### Assistant:"
inputs = tokenizer(
prompt,
return_tensors="pt",
truncation=True,
max_length=512
).to(device)
with torch.no_grad():
outputs = model.generate(
**inputs,
max_new_tokens=400,
temperature=0.7,
top_p=0.9,
do_sample=True,
pad_token_id=tokenizer.eos_token_id,
eos_token_id=end_token_id or tokenizer.eos_token_id,
repetition_penalty=1.05
)
full_response = tokenizer.decode(outputs[0], skip_special_tokens=False)
assistant_response = full_response[len(prompt):].strip()
if "### End" in assistant_response:
assistant_response = assistant_response.split("### End")[0].strip()
assistant_response = re.sub(r'^###\s*Assistant:\s*', '', assistant_response)
if assistant_response:
print(f"\nBot: {assistant_response}")
else:
print("\nBot: please again say")Model Architecture and Objectivepythontransformers
import torch
from transformers import GPT2LMHeadModel, GPT2Tokenizer
import re
model_path = "./Path_To_Model"
print(f"{model_path}...")
tokenizer = GPT2Tokenizer.from_pretrained(model_path)
model = GPT2LMHeadModel.from_pretrained(model_path)
device = "cuda" if torch.cuda.is_available() else "cpu"
print(f"{device}")
model.to(device)
model.eval()
try:
end_token_id = tokenizer.convert_tokens_to_ids("### End")
if end_token_id == tokenizer.unk_token_id:
end_token_id = None
except:
end_token_id = None
print("\n------+------+------+------ model is ready for testing +------+------++------\n")
print("type exit for exit")
print("\nYSNRFD")
print("------+------+------+----------------------+------+------++------\n")
while True:
user_input = input("\nYou: ").strip()
if user_input.lower() in ["exit"]:
print("\n good bye (ysnrfd)")
break
prompt = f"### Human: {user_input}\n### Assistant:"
inputs = tokenizer(
prompt,
return_tensors="pt",
truncation=True,
max_length=512
).to(device)
with torch.no_grad():
outputs = model.generate(
**inputs,
max_new_tokens=400,
temperature=0.7,
top_p=0.9,
do_sample=True,
pad_token_id=tokenizer.eos_token_id,
eos_token_id=end_token_id or tokenizer.eos_token_id,
repetition_penalty=1.05
)
full_response = tokenizer.decode(outputs[0], skip_special_tokens=False)
assistant_response = full_response[len(prompt):].strip()
if "### End" in assistant_response:
assistant_response = assistant_response.split("### End")[0].strip()
assistant_response = re.sub(r'^###\s*Assistant:\s*', '', assistant_response)
if assistant_response:
print(f"\nBot: {assistant_response}")
else:
print("\nBot: please again say")Model Architecture and Objectivepythontransformers
import torch
from transformers import GPT2LMHeadModel, GPT2Tokenizer
import re
model_path = "./Path_To_Model"
print(f"{model_path}...")
tokenizer = GPT2Tokenizer.from_pretrained(model_path)
model = GPT2LMHeadModel.from_pretrained(model_path)
device = "cuda" if torch.cuda.is_available() else "cpu"
print(f"{device}")
model.to(device)
model.eval()
try:
end_token_id = tokenizer.convert_tokens_to_ids("### End")
if end_token_id == tokenizer.unk_token_id:
end_token_id = None
except:
end_token_id = None
print("\n------+------+------+------ model is ready for testing +------+------++------\n")
print("type exit for exit")
print("\nYSNRFD")
print("------+------+------+----------------------+------+------++------\n")
while True:
user_input = input("\nYou: ").strip()
if user_input.lower() in ["exit"]:
print("\n good bye (ysnrfd)")
break
prompt = f"### Human: {user_input}\n### Assistant:"
inputs = tokenizer(
prompt,
return_tensors="pt",
truncation=True,
max_length=512
).to(device)
with torch.no_grad():
outputs = model.generate(
**inputs,
max_new_tokens=400,
temperature=0.7,
top_p=0.9,
do_sample=True,
pad_token_id=tokenizer.eos_token_id,
eos_token_id=end_token_id or tokenizer.eos_token_id,
repetition_penalty=1.05
)
full_response = tokenizer.decode(outputs[0], skip_special_tokens=False)
assistant_response = full_response[len(prompt):].strip()
if "### End" in assistant_response:
assistant_response = assistant_response.split("### End")[0].strip()
assistant_response = re.sub(r'^###\s*Assistant:\s*', '', assistant_response)
if assistant_response:
print(f"\nBot: {assistant_response}")
else:
print("\nBot: please again say")Model Architecture and Objectivepythontransformers
import torch
from transformers import GPT2LMHeadModel, GPT2Tokenizer
import re
model_path = "./Path_To_Model"
print(f"{model_path}...")
tokenizer = GPT2Tokenizer.from_pretrained(model_path)
model = GPT2LMHeadModel.from_pretrained(model_path)
device = "cuda" if torch.cuda.is_available() else "cpu"
print(f"{device}")
model.to(device)
model.eval()
try:
end_token_id = tokenizer.convert_tokens_to_ids("### End")
if end_token_id == tokenizer.unk_token_id:
end_token_id = None
except:
end_token_id = None
print("\n------+------+------+------ model is ready for testing +------+------++------\n")
print("type exit for exit")
print("\nYSNRFD")
print("------+------+------+----------------------+------+------++------\n")
while True:
user_input = input("\nYou: ").strip()
if user_input.lower() in ["exit"]:
print("\n good bye (ysnrfd)")
break
prompt = f"### Human: {user_input}\n### Assistant:"
inputs = tokenizer(
prompt,
return_tensors="pt",
truncation=True,
max_length=512
).to(device)
with torch.no_grad():
outputs = model.generate(
**inputs,
max_new_tokens=400,
temperature=0.7,
top_p=0.9,
do_sample=True,
pad_token_id=tokenizer.eos_token_id,
eos_token_id=end_token_id or tokenizer.eos_token_id,
repetition_penalty=1.05
)
full_response = tokenizer.decode(outputs[0], skip_special_tokens=False)
assistant_response = full_response[len(prompt):].strip()
if "### End" in assistant_response:
assistant_response = assistant_response.split("### End")[0].strip()
assistant_response = re.sub(r'^###\s*Assistant:\s*', '', assistant_response)
if assistant_response:
print(f"\nBot: {assistant_response}")
else:
print("\nBot: please again say")Model Architecture and Objectivepythontransformers
import torch
from transformers import GPT2LMHeadModel, GPT2Tokenizer
import re
model_path = "./Path_To_Model"
print(f"{model_path}...")
tokenizer = GPT2Tokenizer.from_pretrained(model_path)
model = GPT2LMHeadModel.from_pretrained(model_path)
device = "cuda" if torch.cuda.is_available() else "cpu"
print(f"{device}")
model.to(device)
model.eval()
try:
end_token_id = tokenizer.convert_tokens_to_ids("### End")
if end_token_id == tokenizer.unk_token_id:
end_token_id = None
except:
end_token_id = None
print("\n------+------+------+------ model is ready for testing +------+------++------\n")
print("type exit for exit")
print("\nYSNRFD")
print("------+------+------+----------------------+------+------++------\n")
while True:
user_input = input("\nYou: ").strip()
if user_input.lower() in ["exit"]:
print("\n good bye (ysnrfd)")
break
prompt = f"### Human: {user_input}\n### Assistant:"
inputs = tokenizer(
prompt,
return_tensors="pt",
truncation=True,
max_length=512
).to(device)
with torch.no_grad():
outputs = model.generate(
**inputs,
max_new_tokens=400,
temperature=0.7,
top_p=0.9,
do_sample=True,
pad_token_id=tokenizer.eos_token_id,
eos_token_id=end_token_id or tokenizer.eos_token_id,
repetition_penalty=1.05
)
full_response = tokenizer.decode(outputs[0], skip_special_tokens=False)
assistant_response = full_response[len(prompt):].strip()
if "### End" in assistant_response:
assistant_response = assistant_response.split("### End")[0].strip()
assistant_response = re.sub(r'^###\s*Assistant:\s*', '', assistant_response)
if assistant_response:
print(f"\nBot: {assistant_response}")
else:
print("\nBot: please again say")Model Architecture and Objectivepythontransformers
import torch
from transformers import GPT2LMHeadModel, GPT2Tokenizer
import re
model_path = "./Path_To_Model"
print(f"{model_path}...")
tokenizer = GPT2Tokenizer.from_pretrained(model_path)
model = GPT2LMHeadModel.from_pretrained(model_path)
device = "cuda" if torch.cuda.is_available() else "cpu"
print(f"{device}")
model.to(device)
model.eval()
try:
end_token_id = tokenizer.convert_tokens_to_ids("### End")
if end_token_id == tokenizer.unk_token_id:
end_token_id = None
except:
end_token_id = None
print("\n------+------+------+------ model is ready for testing +------+------++------\n")
print("type exit for exit")
print("\nYSNRFD")
print("------+------+------+----------------------+------+------++------\n")
while True:
user_input = input("\nYou: ").strip()
if user_input.lower() in ["exit"]:
print("\n good bye (ysnrfd)")
break
prompt = f"### Human: {user_input}\n### Assistant:"
inputs = tokenizer(
prompt,
return_tensors="pt",
truncation=True,
max_length=512
).to(device)
with torch.no_grad():
outputs = model.generate(
**inputs,
max_new_tokens=400,
temperature=0.7,
top_p=0.9,
do_sample=True,
pad_token_id=tokenizer.eos_token_id,
eos_token_id=end_token_id or tokenizer.eos_token_id,
repetition_penalty=1.05
)
full_response = tokenizer.decode(outputs[0], skip_special_tokens=False)
assistant_response = full_response[len(prompt):].strip()
if "### End" in assistant_response:
assistant_response = assistant_response.split("### End")[0].strip()
assistant_response = re.sub(r'^###\s*Assistant:\s*', '', assistant_response)
if assistant_response:
print(f"\nBot: {assistant_response}")
else:
print("\nBot: please again say")Model Architecture and Objectivepythontransformers
import torch
from transformers import GPT2LMHeadModel, GPT2Tokenizer
import re
model_path = "./Path_To_Model"
print(f"{model_path}...")
tokenizer = GPT2Tokenizer.from_pretrained(model_path)
model = GPT2LMHeadModel.from_pretrained(model_path)
device = "cuda" if torch.cuda.is_available() else "cpu"
print(f"{device}")
model.to(device)
model.eval()
try:
end_token_id = tokenizer.convert_tokens_to_ids("### End")
if end_token_id == tokenizer.unk_token_id:
end_token_id = None
except:
end_token_id = None
print("\n------+------+------+------ model is ready for testing +------+------++------\n")
print("type exit for exit")
print("\nYSNRFD")
print("------+------+------+----------------------+------+------++------\n")
while True:
user_input = input("\nYou: ").strip()
if user_input.lower() in ["exit"]:
print("\n good bye (ysnrfd)")
break
prompt = f"### Human: {user_input}\n### Assistant:"
inputs = tokenizer(
prompt,
return_tensors="pt",
truncation=True,
max_length=512
).to(device)
with torch.no_grad():
outputs = model.generate(
**inputs,
max_new_tokens=400,
temperature=0.7,
top_p=0.9,
do_sample=True,
pad_token_id=tokenizer.eos_token_id,
eos_token_id=end_token_id or tokenizer.eos_token_id,
repetition_penalty=1.05
)
full_response = tokenizer.decode(outputs[0], skip_special_tokens=False)
assistant_response = full_response[len(prompt):].strip()
if "### End" in assistant_response:
assistant_response = assistant_response.split("### End")[0].strip()
assistant_response = re.sub(r'^###\s*Assistant:\s*', '', assistant_response)
if assistant_response:
print(f"\nBot: {assistant_response}")
else:
print("\nBot: please again say")Deploy This Model
Production-ready deployment in minutes
Together.ai
Instant API access to this model
Production-ready inference API. Start free, scale to millions.
Try Free APIReplicate
One-click model deployment
Run models in the cloud with simple API. No DevOps required.
Deploy NowDisclosure: We may earn a commission from these partners. This helps keep LLMYourWay free.