DreamOn-v0-7B
727
17
7.0B
license:apache-2.0
by
Dream-org
Language Model
OTHER
7B params
New
727 downloads
Early-stage
Edge AI:
Mobile
Laptop
Server
16GB+ RAM
Mobile
Laptop
Server
Quick Summary
DreamOn-v0-7B DreamOn is a novel discrete diffusion algorithm designed to address the variable-length generation challenge in code infilling.
Device Compatibility
Mobile
4-6GB RAM
Laptop
16GB RAM
Server
GPU
Minimum Recommended
7GB+ RAM
Code Examples
Quick Starttexttransformers
import torch
import time
from transformers import AutoModel, AutoTokenizer
def process_infilling_prompt(prefix, suffix, tokenizer, number_of_mask):
prefix = [tokenizer.bos_token_id] + tokenizer.encode(prefix, add_special_tokens=False)
middle = [tokenizer.mask_token_id] * number_of_mask
suffix = tokenizer.encode(suffix, add_special_tokens=False) + [tokenizer.eos_token_id]
return prefix + middle + suffix
prefix = '''from typing import List
def has_close_elements(numbers: List[float], threshold: float) -> bool:
""" Check if in given list of numbers, are any two numbers closer to each other than
given threshold.
>>> has_close_elements([1.0, 2.0, 3.0], 0.5)
False
>>> has_close_elements([1.0, 2.8, 3.0, 4.0, 5.0, 2.0], 0.3)
True
"""
'''
suffix = ''' for idx2, elem2 in enumerate(numbers):
if idx != idx2:
distance = abs(elem - elem2)
if distance < threshold:
return True
return False
'''
model_path = 'Dream-org/DreamOn-v0-7B'
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
## Set the initial mask token length when processing the prompt
input_ids = process_infilling_prompt(prefix, suffix, tokenizer, 4)
input_ids = torch.LongTensor([input_ids]).to("cuda")
model = AutoModel.from_pretrained(model_path, torch_dtype=torch.bfloat16, trust_remote_code=True)
model = model.to("cuda").eval()
output = model.diffusion_generate(
input_ids,
temperature=0.2,
alg = 'entropy',
alg_temp = 0,
top_p = 0.9,
max_new_tokens = 64, ## Set the maximum number of new tokens for infilling
return_dict_in_generate = True,
output_history = True,
number_transfer_tokens = 1
)
history = output.history
for i, h in enumerate(history):
print(f"########################")
time.sleep(0.2)
print(tokenizer.decode(h.tolist()), end="\n\n")Quick Starttexttransformers
import torch
import time
from transformers import AutoModel, AutoTokenizer
def process_infilling_prompt(prefix, suffix, tokenizer, number_of_mask):
prefix = [tokenizer.bos_token_id] + tokenizer.encode(prefix, add_special_tokens=False)
middle = [tokenizer.mask_token_id] * number_of_mask
suffix = tokenizer.encode(suffix, add_special_tokens=False) + [tokenizer.eos_token_id]
return prefix + middle + suffix
prefix = '''from typing import List
def has_close_elements(numbers: List[float], threshold: float) -> bool:
""" Check if in given list of numbers, are any two numbers closer to each other than
given threshold.
>>> has_close_elements([1.0, 2.0, 3.0], 0.5)
False
>>> has_close_elements([1.0, 2.8, 3.0, 4.0, 5.0, 2.0], 0.3)
True
"""
'''
suffix = ''' for idx2, elem2 in enumerate(numbers):
if idx != idx2:
distance = abs(elem - elem2)
if distance < threshold:
return True
return False
'''
model_path = 'Dream-org/DreamOn-v0-7B'
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
## Set the initial mask token length when processing the prompt
input_ids = process_infilling_prompt(prefix, suffix, tokenizer, 4)
input_ids = torch.LongTensor([input_ids]).to("cuda")
model = AutoModel.from_pretrained(model_path, torch_dtype=torch.bfloat16, trust_remote_code=True)
model = model.to("cuda").eval()
output = model.diffusion_generate(
input_ids,
temperature=0.2,
alg = 'entropy',
alg_temp = 0,
top_p = 0.9,
max_new_tokens = 64, ## Set the maximum number of new tokens for infilling
return_dict_in_generate = True,
output_history = True,
number_transfer_tokens = 1
)
history = output.history
for i, h in enumerate(history):
print(f"########################")
time.sleep(0.2)
print(tokenizer.decode(h.tolist()), end="\n\n")Quick Starttexttransformers
import torch
import time
from transformers import AutoModel, AutoTokenizer
def process_infilling_prompt(prefix, suffix, tokenizer, number_of_mask):
prefix = [tokenizer.bos_token_id] + tokenizer.encode(prefix, add_special_tokens=False)
middle = [tokenizer.mask_token_id] * number_of_mask
suffix = tokenizer.encode(suffix, add_special_tokens=False) + [tokenizer.eos_token_id]
return prefix + middle + suffix
prefix = '''from typing import List
def has_close_elements(numbers: List[float], threshold: float) -> bool:
""" Check if in given list of numbers, are any two numbers closer to each other than
given threshold.
>>> has_close_elements([1.0, 2.0, 3.0], 0.5)
False
>>> has_close_elements([1.0, 2.8, 3.0, 4.0, 5.0, 2.0], 0.3)
True
"""
'''
suffix = ''' for idx2, elem2 in enumerate(numbers):
if idx != idx2:
distance = abs(elem - elem2)
if distance < threshold:
return True
return False
'''
model_path = 'Dream-org/DreamOn-v0-7B'
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
## Set the initial mask token length when processing the prompt
input_ids = process_infilling_prompt(prefix, suffix, tokenizer, 4)
input_ids = torch.LongTensor([input_ids]).to("cuda")
model = AutoModel.from_pretrained(model_path, torch_dtype=torch.bfloat16, trust_remote_code=True)
model = model.to("cuda").eval()
output = model.diffusion_generate(
input_ids,
temperature=0.2,
alg = 'entropy',
alg_temp = 0,
top_p = 0.9,
max_new_tokens = 64, ## Set the maximum number of new tokens for infilling
return_dict_in_generate = True,
output_history = True,
number_transfer_tokens = 1
)
history = output.history
for i, h in enumerate(history):
print(f"########################")
time.sleep(0.2)
print(tokenizer.decode(h.tolist()), end="\n\n")Quick Starttexttransformers
import torch
import time
from transformers import AutoModel, AutoTokenizer
def process_infilling_prompt(prefix, suffix, tokenizer, number_of_mask):
prefix = [tokenizer.bos_token_id] + tokenizer.encode(prefix, add_special_tokens=False)
middle = [tokenizer.mask_token_id] * number_of_mask
suffix = tokenizer.encode(suffix, add_special_tokens=False) + [tokenizer.eos_token_id]
return prefix + middle + suffix
prefix = '''from typing import List
def has_close_elements(numbers: List[float], threshold: float) -> bool:
""" Check if in given list of numbers, are any two numbers closer to each other than
given threshold.
>>> has_close_elements([1.0, 2.0, 3.0], 0.5)
False
>>> has_close_elements([1.0, 2.8, 3.0, 4.0, 5.0, 2.0], 0.3)
True
"""
'''
suffix = ''' for idx2, elem2 in enumerate(numbers):
if idx != idx2:
distance = abs(elem - elem2)
if distance < threshold:
return True
return False
'''
model_path = 'Dream-org/DreamOn-v0-7B'
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
## Set the initial mask token length when processing the prompt
input_ids = process_infilling_prompt(prefix, suffix, tokenizer, 4)
input_ids = torch.LongTensor([input_ids]).to("cuda")
model = AutoModel.from_pretrained(model_path, torch_dtype=torch.bfloat16, trust_remote_code=True)
model = model.to("cuda").eval()
output = model.diffusion_generate(
input_ids,
temperature=0.2,
alg = 'entropy',
alg_temp = 0,
top_p = 0.9,
max_new_tokens = 64, ## Set the maximum number of new tokens for infilling
return_dict_in_generate = True,
output_history = True,
number_transfer_tokens = 1
)
history = output.history
for i, h in enumerate(history):
print(f"########################")
time.sleep(0.2)
print(tokenizer.decode(h.tolist()), end="\n\n")Quick Starttexttransformers
import torch
import time
from transformers import AutoModel, AutoTokenizer
def process_infilling_prompt(prefix, suffix, tokenizer, number_of_mask):
prefix = [tokenizer.bos_token_id] + tokenizer.encode(prefix, add_special_tokens=False)
middle = [tokenizer.mask_token_id] * number_of_mask
suffix = tokenizer.encode(suffix, add_special_tokens=False) + [tokenizer.eos_token_id]
return prefix + middle + suffix
prefix = '''from typing import List
def has_close_elements(numbers: List[float], threshold: float) -> bool:
""" Check if in given list of numbers, are any two numbers closer to each other than
given threshold.
>>> has_close_elements([1.0, 2.0, 3.0], 0.5)
False
>>> has_close_elements([1.0, 2.8, 3.0, 4.0, 5.0, 2.0], 0.3)
True
"""
'''
suffix = ''' for idx2, elem2 in enumerate(numbers):
if idx != idx2:
distance = abs(elem - elem2)
if distance < threshold:
return True
return False
'''
model_path = 'Dream-org/DreamOn-v0-7B'
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
## Set the initial mask token length when processing the prompt
input_ids = process_infilling_prompt(prefix, suffix, tokenizer, 4)
input_ids = torch.LongTensor([input_ids]).to("cuda")
model = AutoModel.from_pretrained(model_path, torch_dtype=torch.bfloat16, trust_remote_code=True)
model = model.to("cuda").eval()
output = model.diffusion_generate(
input_ids,
temperature=0.2,
alg = 'entropy',
alg_temp = 0,
top_p = 0.9,
max_new_tokens = 64, ## Set the maximum number of new tokens for infilling
return_dict_in_generate = True,
output_history = True,
number_transfer_tokens = 1
)
history = output.history
for i, h in enumerate(history):
print(f"########################")
time.sleep(0.2)
print(tokenizer.decode(h.tolist()), end="\n\n")Quick Starttexttransformers
import torch
import time
from transformers import AutoModel, AutoTokenizer
def process_infilling_prompt(prefix, suffix, tokenizer, number_of_mask):
prefix = [tokenizer.bos_token_id] + tokenizer.encode(prefix, add_special_tokens=False)
middle = [tokenizer.mask_token_id] * number_of_mask
suffix = tokenizer.encode(suffix, add_special_tokens=False) + [tokenizer.eos_token_id]
return prefix + middle + suffix
prefix = '''from typing import List
def has_close_elements(numbers: List[float], threshold: float) -> bool:
""" Check if in given list of numbers, are any two numbers closer to each other than
given threshold.
>>> has_close_elements([1.0, 2.0, 3.0], 0.5)
False
>>> has_close_elements([1.0, 2.8, 3.0, 4.0, 5.0, 2.0], 0.3)
True
"""
'''
suffix = ''' for idx2, elem2 in enumerate(numbers):
if idx != idx2:
distance = abs(elem - elem2)
if distance < threshold:
return True
return False
'''
model_path = 'Dream-org/DreamOn-v0-7B'
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
## Set the initial mask token length when processing the prompt
input_ids = process_infilling_prompt(prefix, suffix, tokenizer, 4)
input_ids = torch.LongTensor([input_ids]).to("cuda")
model = AutoModel.from_pretrained(model_path, torch_dtype=torch.bfloat16, trust_remote_code=True)
model = model.to("cuda").eval()
output = model.diffusion_generate(
input_ids,
temperature=0.2,
alg = 'entropy',
alg_temp = 0,
top_p = 0.9,
max_new_tokens = 64, ## Set the maximum number of new tokens for infilling
return_dict_in_generate = True,
output_history = True,
number_transfer_tokens = 1
)
history = output.history
for i, h in enumerate(history):
print(f"########################")
time.sleep(0.2)
print(tokenizer.decode(h.tolist()), end="\n\n")Quick Starttexttransformers
import torch
import time
from transformers import AutoModel, AutoTokenizer
def process_infilling_prompt(prefix, suffix, tokenizer, number_of_mask):
prefix = [tokenizer.bos_token_id] + tokenizer.encode(prefix, add_special_tokens=False)
middle = [tokenizer.mask_token_id] * number_of_mask
suffix = tokenizer.encode(suffix, add_special_tokens=False) + [tokenizer.eos_token_id]
return prefix + middle + suffix
prefix = '''from typing import List
def has_close_elements(numbers: List[float], threshold: float) -> bool:
""" Check if in given list of numbers, are any two numbers closer to each other than
given threshold.
>>> has_close_elements([1.0, 2.0, 3.0], 0.5)
False
>>> has_close_elements([1.0, 2.8, 3.0, 4.0, 5.0, 2.0], 0.3)
True
"""
'''
suffix = ''' for idx2, elem2 in enumerate(numbers):
if idx != idx2:
distance = abs(elem - elem2)
if distance < threshold:
return True
return False
'''
model_path = 'Dream-org/DreamOn-v0-7B'
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
## Set the initial mask token length when processing the prompt
input_ids = process_infilling_prompt(prefix, suffix, tokenizer, 4)
input_ids = torch.LongTensor([input_ids]).to("cuda")
model = AutoModel.from_pretrained(model_path, torch_dtype=torch.bfloat16, trust_remote_code=True)
model = model.to("cuda").eval()
output = model.diffusion_generate(
input_ids,
temperature=0.2,
alg = 'entropy',
alg_temp = 0,
top_p = 0.9,
max_new_tokens = 64, ## Set the maximum number of new tokens for infilling
return_dict_in_generate = True,
output_history = True,
number_transfer_tokens = 1
)
history = output.history
for i, h in enumerate(history):
print(f"########################")
time.sleep(0.2)
print(tokenizer.decode(h.tolist()), end="\n\n")Quick Starttexttransformers
import torch
import time
from transformers import AutoModel, AutoTokenizer
def process_infilling_prompt(prefix, suffix, tokenizer, number_of_mask):
prefix = [tokenizer.bos_token_id] + tokenizer.encode(prefix, add_special_tokens=False)
middle = [tokenizer.mask_token_id] * number_of_mask
suffix = tokenizer.encode(suffix, add_special_tokens=False) + [tokenizer.eos_token_id]
return prefix + middle + suffix
prefix = '''from typing import List
def has_close_elements(numbers: List[float], threshold: float) -> bool:
""" Check if in given list of numbers, are any two numbers closer to each other than
given threshold.
>>> has_close_elements([1.0, 2.0, 3.0], 0.5)
False
>>> has_close_elements([1.0, 2.8, 3.0, 4.0, 5.0, 2.0], 0.3)
True
"""
'''
suffix = ''' for idx2, elem2 in enumerate(numbers):
if idx != idx2:
distance = abs(elem - elem2)
if distance < threshold:
return True
return False
'''
model_path = 'Dream-org/DreamOn-v0-7B'
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
## Set the initial mask token length when processing the prompt
input_ids = process_infilling_prompt(prefix, suffix, tokenizer, 4)
input_ids = torch.LongTensor([input_ids]).to("cuda")
model = AutoModel.from_pretrained(model_path, torch_dtype=torch.bfloat16, trust_remote_code=True)
model = model.to("cuda").eval()
output = model.diffusion_generate(
input_ids,
temperature=0.2,
alg = 'entropy',
alg_temp = 0,
top_p = 0.9,
max_new_tokens = 64, ## Set the maximum number of new tokens for infilling
return_dict_in_generate = True,
output_history = True,
number_transfer_tokens = 1
)
history = output.history
for i, h in enumerate(history):
print(f"########################")
time.sleep(0.2)
print(tokenizer.decode(h.tolist()), end="\n\n")Quick Starttexttransformers
import torch
import time
from transformers import AutoModel, AutoTokenizer
def process_infilling_prompt(prefix, suffix, tokenizer, number_of_mask):
prefix = [tokenizer.bos_token_id] + tokenizer.encode(prefix, add_special_tokens=False)
middle = [tokenizer.mask_token_id] * number_of_mask
suffix = tokenizer.encode(suffix, add_special_tokens=False) + [tokenizer.eos_token_id]
return prefix + middle + suffix
prefix = '''from typing import List
def has_close_elements(numbers: List[float], threshold: float) -> bool:
""" Check if in given list of numbers, are any two numbers closer to each other than
given threshold.
>>> has_close_elements([1.0, 2.0, 3.0], 0.5)
False
>>> has_close_elements([1.0, 2.8, 3.0, 4.0, 5.0, 2.0], 0.3)
True
"""
'''
suffix = ''' for idx2, elem2 in enumerate(numbers):
if idx != idx2:
distance = abs(elem - elem2)
if distance < threshold:
return True
return False
'''
model_path = 'Dream-org/DreamOn-v0-7B'
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
## Set the initial mask token length when processing the prompt
input_ids = process_infilling_prompt(prefix, suffix, tokenizer, 4)
input_ids = torch.LongTensor([input_ids]).to("cuda")
model = AutoModel.from_pretrained(model_path, torch_dtype=torch.bfloat16, trust_remote_code=True)
model = model.to("cuda").eval()
output = model.diffusion_generate(
input_ids,
temperature=0.2,
alg = 'entropy',
alg_temp = 0,
top_p = 0.9,
max_new_tokens = 64, ## Set the maximum number of new tokens for infilling
return_dict_in_generate = True,
output_history = True,
number_transfer_tokens = 1
)
history = output.history
for i, h in enumerate(history):
print(f"########################")
time.sleep(0.2)
print(tokenizer.decode(h.tolist()), end="\n\n")Quick Starttexttransformers
import torch
import time
from transformers import AutoModel, AutoTokenizer
def process_infilling_prompt(prefix, suffix, tokenizer, number_of_mask):
prefix = [tokenizer.bos_token_id] + tokenizer.encode(prefix, add_special_tokens=False)
middle = [tokenizer.mask_token_id] * number_of_mask
suffix = tokenizer.encode(suffix, add_special_tokens=False) + [tokenizer.eos_token_id]
return prefix + middle + suffix
prefix = '''from typing import List
def has_close_elements(numbers: List[float], threshold: float) -> bool:
""" Check if in given list of numbers, are any two numbers closer to each other than
given threshold.
>>> has_close_elements([1.0, 2.0, 3.0], 0.5)
False
>>> has_close_elements([1.0, 2.8, 3.0, 4.0, 5.0, 2.0], 0.3)
True
"""
'''
suffix = ''' for idx2, elem2 in enumerate(numbers):
if idx != idx2:
distance = abs(elem - elem2)
if distance < threshold:
return True
return False
'''
model_path = 'Dream-org/DreamOn-v0-7B'
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
## Set the initial mask token length when processing the prompt
input_ids = process_infilling_prompt(prefix, suffix, tokenizer, 4)
input_ids = torch.LongTensor([input_ids]).to("cuda")
model = AutoModel.from_pretrained(model_path, torch_dtype=torch.bfloat16, trust_remote_code=True)
model = model.to("cuda").eval()
output = model.diffusion_generate(
input_ids,
temperature=0.2,
alg = 'entropy',
alg_temp = 0,
top_p = 0.9,
max_new_tokens = 64, ## Set the maximum number of new tokens for infilling
return_dict_in_generate = True,
output_history = True,
number_transfer_tokens = 1
)
history = output.history
for i, h in enumerate(history):
print(f"########################")
time.sleep(0.2)
print(tokenizer.decode(h.tolist()), end="\n\n")Quick Starttexttransformers
import torch
import time
from transformers import AutoModel, AutoTokenizer
def process_infilling_prompt(prefix, suffix, tokenizer, number_of_mask):
prefix = [tokenizer.bos_token_id] + tokenizer.encode(prefix, add_special_tokens=False)
middle = [tokenizer.mask_token_id] * number_of_mask
suffix = tokenizer.encode(suffix, add_special_tokens=False) + [tokenizer.eos_token_id]
return prefix + middle + suffix
prefix = '''from typing import List
def has_close_elements(numbers: List[float], threshold: float) -> bool:
""" Check if in given list of numbers, are any two numbers closer to each other than
given threshold.
>>> has_close_elements([1.0, 2.0, 3.0], 0.5)
False
>>> has_close_elements([1.0, 2.8, 3.0, 4.0, 5.0, 2.0], 0.3)
True
"""
'''
suffix = ''' for idx2, elem2 in enumerate(numbers):
if idx != idx2:
distance = abs(elem - elem2)
if distance < threshold:
return True
return False
'''
model_path = 'Dream-org/DreamOn-v0-7B'
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
## Set the initial mask token length when processing the prompt
input_ids = process_infilling_prompt(prefix, suffix, tokenizer, 4)
input_ids = torch.LongTensor([input_ids]).to("cuda")
model = AutoModel.from_pretrained(model_path, torch_dtype=torch.bfloat16, trust_remote_code=True)
model = model.to("cuda").eval()
output = model.diffusion_generate(
input_ids,
temperature=0.2,
alg = 'entropy',
alg_temp = 0,
top_p = 0.9,
max_new_tokens = 64, ## Set the maximum number of new tokens for infilling
return_dict_in_generate = True,
output_history = True,
number_transfer_tokens = 1
)
history = output.history
for i, h in enumerate(history):
print(f"########################")
time.sleep(0.2)
print(tokenizer.decode(h.tolist()), end="\n\n")Quick Starttexttransformers
import torch
import time
from transformers import AutoModel, AutoTokenizer
def process_infilling_prompt(prefix, suffix, tokenizer, number_of_mask):
prefix = [tokenizer.bos_token_id] + tokenizer.encode(prefix, add_special_tokens=False)
middle = [tokenizer.mask_token_id] * number_of_mask
suffix = tokenizer.encode(suffix, add_special_tokens=False) + [tokenizer.eos_token_id]
return prefix + middle + suffix
prefix = '''from typing import List
def has_close_elements(numbers: List[float], threshold: float) -> bool:
""" Check if in given list of numbers, are any two numbers closer to each other than
given threshold.
>>> has_close_elements([1.0, 2.0, 3.0], 0.5)
False
>>> has_close_elements([1.0, 2.8, 3.0, 4.0, 5.0, 2.0], 0.3)
True
"""
'''
suffix = ''' for idx2, elem2 in enumerate(numbers):
if idx != idx2:
distance = abs(elem - elem2)
if distance < threshold:
return True
return False
'''
model_path = 'Dream-org/DreamOn-v0-7B'
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
## Set the initial mask token length when processing the prompt
input_ids = process_infilling_prompt(prefix, suffix, tokenizer, 4)
input_ids = torch.LongTensor([input_ids]).to("cuda")
model = AutoModel.from_pretrained(model_path, torch_dtype=torch.bfloat16, trust_remote_code=True)
model = model.to("cuda").eval()
output = model.diffusion_generate(
input_ids,
temperature=0.2,
alg = 'entropy',
alg_temp = 0,
top_p = 0.9,
max_new_tokens = 64, ## Set the maximum number of new tokens for infilling
return_dict_in_generate = True,
output_history = True,
number_transfer_tokens = 1
)
history = output.history
for i, h in enumerate(history):
print(f"########################")
time.sleep(0.2)
print(tokenizer.decode(h.tolist()), end="\n\n")Quick Starttexttransformers
import torch
import time
from transformers import AutoModel, AutoTokenizer
def process_infilling_prompt(prefix, suffix, tokenizer, number_of_mask):
prefix = [tokenizer.bos_token_id] + tokenizer.encode(prefix, add_special_tokens=False)
middle = [tokenizer.mask_token_id] * number_of_mask
suffix = tokenizer.encode(suffix, add_special_tokens=False) + [tokenizer.eos_token_id]
return prefix + middle + suffix
prefix = '''from typing import List
def has_close_elements(numbers: List[float], threshold: float) -> bool:
""" Check if in given list of numbers, are any two numbers closer to each other than
given threshold.
>>> has_close_elements([1.0, 2.0, 3.0], 0.5)
False
>>> has_close_elements([1.0, 2.8, 3.0, 4.0, 5.0, 2.0], 0.3)
True
"""
'''
suffix = ''' for idx2, elem2 in enumerate(numbers):
if idx != idx2:
distance = abs(elem - elem2)
if distance < threshold:
return True
return False
'''
model_path = 'Dream-org/DreamOn-v0-7B'
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
## Set the initial mask token length when processing the prompt
input_ids = process_infilling_prompt(prefix, suffix, tokenizer, 4)
input_ids = torch.LongTensor([input_ids]).to("cuda")
model = AutoModel.from_pretrained(model_path, torch_dtype=torch.bfloat16, trust_remote_code=True)
model = model.to("cuda").eval()
output = model.diffusion_generate(
input_ids,
temperature=0.2,
alg = 'entropy',
alg_temp = 0,
top_p = 0.9,
max_new_tokens = 64, ## Set the maximum number of new tokens for infilling
return_dict_in_generate = True,
output_history = True,
number_transfer_tokens = 1
)
history = output.history
for i, h in enumerate(history):
print(f"########################")
time.sleep(0.2)
print(tokenizer.decode(h.tolist()), end="\n\n")Quick Starttexttransformers
import torch
import time
from transformers import AutoModel, AutoTokenizer
def process_infilling_prompt(prefix, suffix, tokenizer, number_of_mask):
prefix = [tokenizer.bos_token_id] + tokenizer.encode(prefix, add_special_tokens=False)
middle = [tokenizer.mask_token_id] * number_of_mask
suffix = tokenizer.encode(suffix, add_special_tokens=False) + [tokenizer.eos_token_id]
return prefix + middle + suffix
prefix = '''from typing import List
def has_close_elements(numbers: List[float], threshold: float) -> bool:
""" Check if in given list of numbers, are any two numbers closer to each other than
given threshold.
>>> has_close_elements([1.0, 2.0, 3.0], 0.5)
False
>>> has_close_elements([1.0, 2.8, 3.0, 4.0, 5.0, 2.0], 0.3)
True
"""
'''
suffix = ''' for idx2, elem2 in enumerate(numbers):
if idx != idx2:
distance = abs(elem - elem2)
if distance < threshold:
return True
return False
'''
model_path = 'Dream-org/DreamOn-v0-7B'
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
## Set the initial mask token length when processing the prompt
input_ids = process_infilling_prompt(prefix, suffix, tokenizer, 4)
input_ids = torch.LongTensor([input_ids]).to("cuda")
model = AutoModel.from_pretrained(model_path, torch_dtype=torch.bfloat16, trust_remote_code=True)
model = model.to("cuda").eval()
output = model.diffusion_generate(
input_ids,
temperature=0.2,
alg = 'entropy',
alg_temp = 0,
top_p = 0.9,
max_new_tokens = 64, ## Set the maximum number of new tokens for infilling
return_dict_in_generate = True,
output_history = True,
number_transfer_tokens = 1
)
history = output.history
for i, h in enumerate(history):
print(f"########################")
time.sleep(0.2)
print(tokenizer.decode(h.tolist()), end="\n\n")Quick Starttexttransformers
import torch
import time
from transformers import AutoModel, AutoTokenizer
def process_infilling_prompt(prefix, suffix, tokenizer, number_of_mask):
prefix = [tokenizer.bos_token_id] + tokenizer.encode(prefix, add_special_tokens=False)
middle = [tokenizer.mask_token_id] * number_of_mask
suffix = tokenizer.encode(suffix, add_special_tokens=False) + [tokenizer.eos_token_id]
return prefix + middle + suffix
prefix = '''from typing import List
def has_close_elements(numbers: List[float], threshold: float) -> bool:
""" Check if in given list of numbers, are any two numbers closer to each other than
given threshold.
>>> has_close_elements([1.0, 2.0, 3.0], 0.5)
False
>>> has_close_elements([1.0, 2.8, 3.0, 4.0, 5.0, 2.0], 0.3)
True
"""
'''
suffix = ''' for idx2, elem2 in enumerate(numbers):
if idx != idx2:
distance = abs(elem - elem2)
if distance < threshold:
return True
return False
'''
model_path = 'Dream-org/DreamOn-v0-7B'
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
## Set the initial mask token length when processing the prompt
input_ids = process_infilling_prompt(prefix, suffix, tokenizer, 4)
input_ids = torch.LongTensor([input_ids]).to("cuda")
model = AutoModel.from_pretrained(model_path, torch_dtype=torch.bfloat16, trust_remote_code=True)
model = model.to("cuda").eval()
output = model.diffusion_generate(
input_ids,
temperature=0.2,
alg = 'entropy',
alg_temp = 0,
top_p = 0.9,
max_new_tokens = 64, ## Set the maximum number of new tokens for infilling
return_dict_in_generate = True,
output_history = True,
number_transfer_tokens = 1
)
history = output.history
for i, h in enumerate(history):
print(f"########################")
time.sleep(0.2)
print(tokenizer.decode(h.tolist()), end="\n\n")Quick Starttexttransformers
import torch
import time
from transformers import AutoModel, AutoTokenizer
def process_infilling_prompt(prefix, suffix, tokenizer, number_of_mask):
prefix = [tokenizer.bos_token_id] + tokenizer.encode(prefix, add_special_tokens=False)
middle = [tokenizer.mask_token_id] * number_of_mask
suffix = tokenizer.encode(suffix, add_special_tokens=False) + [tokenizer.eos_token_id]
return prefix + middle + suffix
prefix = '''from typing import List
def has_close_elements(numbers: List[float], threshold: float) -> bool:
""" Check if in given list of numbers, are any two numbers closer to each other than
given threshold.
>>> has_close_elements([1.0, 2.0, 3.0], 0.5)
False
>>> has_close_elements([1.0, 2.8, 3.0, 4.0, 5.0, 2.0], 0.3)
True
"""
'''
suffix = ''' for idx2, elem2 in enumerate(numbers):
if idx != idx2:
distance = abs(elem - elem2)
if distance < threshold:
return True
return False
'''
model_path = 'Dream-org/DreamOn-v0-7B'
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
## Set the initial mask token length when processing the prompt
input_ids = process_infilling_prompt(prefix, suffix, tokenizer, 4)
input_ids = torch.LongTensor([input_ids]).to("cuda")
model = AutoModel.from_pretrained(model_path, torch_dtype=torch.bfloat16, trust_remote_code=True)
model = model.to("cuda").eval()
output = model.diffusion_generate(
input_ids,
temperature=0.2,
alg = 'entropy',
alg_temp = 0,
top_p = 0.9,
max_new_tokens = 64, ## Set the maximum number of new tokens for infilling
return_dict_in_generate = True,
output_history = True,
number_transfer_tokens = 1
)
history = output.history
for i, h in enumerate(history):
print(f"########################")
time.sleep(0.2)
print(tokenizer.decode(h.tolist()), end="\n\n")Quick Starttexttransformers
import torch
import time
from transformers import AutoModel, AutoTokenizer
def process_infilling_prompt(prefix, suffix, tokenizer, number_of_mask):
prefix = [tokenizer.bos_token_id] + tokenizer.encode(prefix, add_special_tokens=False)
middle = [tokenizer.mask_token_id] * number_of_mask
suffix = tokenizer.encode(suffix, add_special_tokens=False) + [tokenizer.eos_token_id]
return prefix + middle + suffix
prefix = '''from typing import List
def has_close_elements(numbers: List[float], threshold: float) -> bool:
""" Check if in given list of numbers, are any two numbers closer to each other than
given threshold.
>>> has_close_elements([1.0, 2.0, 3.0], 0.5)
False
>>> has_close_elements([1.0, 2.8, 3.0, 4.0, 5.0, 2.0], 0.3)
True
"""
'''
suffix = ''' for idx2, elem2 in enumerate(numbers):
if idx != idx2:
distance = abs(elem - elem2)
if distance < threshold:
return True
return False
'''
model_path = 'Dream-org/DreamOn-v0-7B'
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
## Set the initial mask token length when processing the prompt
input_ids = process_infilling_prompt(prefix, suffix, tokenizer, 4)
input_ids = torch.LongTensor([input_ids]).to("cuda")
model = AutoModel.from_pretrained(model_path, torch_dtype=torch.bfloat16, trust_remote_code=True)
model = model.to("cuda").eval()
output = model.diffusion_generate(
input_ids,
temperature=0.2,
alg = 'entropy',
alg_temp = 0,
top_p = 0.9,
max_new_tokens = 64, ## Set the maximum number of new tokens for infilling
return_dict_in_generate = True,
output_history = True,
number_transfer_tokens = 1
)
history = output.history
for i, h in enumerate(history):
print(f"########################")
time.sleep(0.2)
print(tokenizer.decode(h.tolist()), end="\n\n")Quick Starttexttransformers
import torch
import time
from transformers import AutoModel, AutoTokenizer
def process_infilling_prompt(prefix, suffix, tokenizer, number_of_mask):
prefix = [tokenizer.bos_token_id] + tokenizer.encode(prefix, add_special_tokens=False)
middle = [tokenizer.mask_token_id] * number_of_mask
suffix = tokenizer.encode(suffix, add_special_tokens=False) + [tokenizer.eos_token_id]
return prefix + middle + suffix
prefix = '''from typing import List
def has_close_elements(numbers: List[float], threshold: float) -> bool:
""" Check if in given list of numbers, are any two numbers closer to each other than
given threshold.
>>> has_close_elements([1.0, 2.0, 3.0], 0.5)
False
>>> has_close_elements([1.0, 2.8, 3.0, 4.0, 5.0, 2.0], 0.3)
True
"""
'''
suffix = ''' for idx2, elem2 in enumerate(numbers):
if idx != idx2:
distance = abs(elem - elem2)
if distance < threshold:
return True
return False
'''
model_path = 'Dream-org/DreamOn-v0-7B'
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
## Set the initial mask token length when processing the prompt
input_ids = process_infilling_prompt(prefix, suffix, tokenizer, 4)
input_ids = torch.LongTensor([input_ids]).to("cuda")
model = AutoModel.from_pretrained(model_path, torch_dtype=torch.bfloat16, trust_remote_code=True)
model = model.to("cuda").eval()
output = model.diffusion_generate(
input_ids,
temperature=0.2,
alg = 'entropy',
alg_temp = 0,
top_p = 0.9,
max_new_tokens = 64, ## Set the maximum number of new tokens for infilling
return_dict_in_generate = True,
output_history = True,
number_transfer_tokens = 1
)
history = output.history
for i, h in enumerate(history):
print(f"########################")
time.sleep(0.2)
print(tokenizer.decode(h.tolist()), end="\n\n")Quick Starttexttransformers
import torch
import time
from transformers import AutoModel, AutoTokenizer
def process_infilling_prompt(prefix, suffix, tokenizer, number_of_mask):
prefix = [tokenizer.bos_token_id] + tokenizer.encode(prefix, add_special_tokens=False)
middle = [tokenizer.mask_token_id] * number_of_mask
suffix = tokenizer.encode(suffix, add_special_tokens=False) + [tokenizer.eos_token_id]
return prefix + middle + suffix
prefix = '''from typing import List
def has_close_elements(numbers: List[float], threshold: float) -> bool:
""" Check if in given list of numbers, are any two numbers closer to each other than
given threshold.
>>> has_close_elements([1.0, 2.0, 3.0], 0.5)
False
>>> has_close_elements([1.0, 2.8, 3.0, 4.0, 5.0, 2.0], 0.3)
True
"""
'''
suffix = ''' for idx2, elem2 in enumerate(numbers):
if idx != idx2:
distance = abs(elem - elem2)
if distance < threshold:
return True
return False
'''
model_path = 'Dream-org/DreamOn-v0-7B'
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
## Set the initial mask token length when processing the prompt
input_ids = process_infilling_prompt(prefix, suffix, tokenizer, 4)
input_ids = torch.LongTensor([input_ids]).to("cuda")
model = AutoModel.from_pretrained(model_path, torch_dtype=torch.bfloat16, trust_remote_code=True)
model = model.to("cuda").eval()
output = model.diffusion_generate(
input_ids,
temperature=0.2,
alg = 'entropy',
alg_temp = 0,
top_p = 0.9,
max_new_tokens = 64, ## Set the maximum number of new tokens for infilling
return_dict_in_generate = True,
output_history = True,
number_transfer_tokens = 1
)
history = output.history
for i, h in enumerate(history):
print(f"########################")
time.sleep(0.2)
print(tokenizer.decode(h.tolist()), end="\n\n")Quick Starttexttransformers
import torch
import time
from transformers import AutoModel, AutoTokenizer
def process_infilling_prompt(prefix, suffix, tokenizer, number_of_mask):
prefix = [tokenizer.bos_token_id] + tokenizer.encode(prefix, add_special_tokens=False)
middle = [tokenizer.mask_token_id] * number_of_mask
suffix = tokenizer.encode(suffix, add_special_tokens=False) + [tokenizer.eos_token_id]
return prefix + middle + suffix
prefix = '''from typing import List
def has_close_elements(numbers: List[float], threshold: float) -> bool:
""" Check if in given list of numbers, are any two numbers closer to each other than
given threshold.
>>> has_close_elements([1.0, 2.0, 3.0], 0.5)
False
>>> has_close_elements([1.0, 2.8, 3.0, 4.0, 5.0, 2.0], 0.3)
True
"""
'''
suffix = ''' for idx2, elem2 in enumerate(numbers):
if idx != idx2:
distance = abs(elem - elem2)
if distance < threshold:
return True
return False
'''
model_path = 'Dream-org/DreamOn-v0-7B'
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
## Set the initial mask token length when processing the prompt
input_ids = process_infilling_prompt(prefix, suffix, tokenizer, 4)
input_ids = torch.LongTensor([input_ids]).to("cuda")
model = AutoModel.from_pretrained(model_path, torch_dtype=torch.bfloat16, trust_remote_code=True)
model = model.to("cuda").eval()
output = model.diffusion_generate(
input_ids,
temperature=0.2,
alg = 'entropy',
alg_temp = 0,
top_p = 0.9,
max_new_tokens = 64, ## Set the maximum number of new tokens for infilling
return_dict_in_generate = True,
output_history = True,
number_transfer_tokens = 1
)
history = output.history
for i, h in enumerate(history):
print(f"########################")
time.sleep(0.2)
print(tokenizer.decode(h.tolist()), end="\n\n")Quick Starttexttransformers
import torch
import time
from transformers import AutoModel, AutoTokenizer
def process_infilling_prompt(prefix, suffix, tokenizer, number_of_mask):
prefix = [tokenizer.bos_token_id] + tokenizer.encode(prefix, add_special_tokens=False)
middle = [tokenizer.mask_token_id] * number_of_mask
suffix = tokenizer.encode(suffix, add_special_tokens=False) + [tokenizer.eos_token_id]
return prefix + middle + suffix
prefix = '''from typing import List
def has_close_elements(numbers: List[float], threshold: float) -> bool:
""" Check if in given list of numbers, are any two numbers closer to each other than
given threshold.
>>> has_close_elements([1.0, 2.0, 3.0], 0.5)
False
>>> has_close_elements([1.0, 2.8, 3.0, 4.0, 5.0, 2.0], 0.3)
True
"""
'''
suffix = ''' for idx2, elem2 in enumerate(numbers):
if idx != idx2:
distance = abs(elem - elem2)
if distance < threshold:
return True
return False
'''
model_path = 'Dream-org/DreamOn-v0-7B'
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
## Set the initial mask token length when processing the prompt
input_ids = process_infilling_prompt(prefix, suffix, tokenizer, 4)
input_ids = torch.LongTensor([input_ids]).to("cuda")
model = AutoModel.from_pretrained(model_path, torch_dtype=torch.bfloat16, trust_remote_code=True)
model = model.to("cuda").eval()
output = model.diffusion_generate(
input_ids,
temperature=0.2,
alg = 'entropy',
alg_temp = 0,
top_p = 0.9,
max_new_tokens = 64, ## Set the maximum number of new tokens for infilling
return_dict_in_generate = True,
output_history = True,
number_transfer_tokens = 1
)
history = output.history
for i, h in enumerate(history):
print(f"########################")
time.sleep(0.2)
print(tokenizer.decode(h.tolist()), end="\n\n")Quick Starttexttransformers
import torch
import time
from transformers import AutoModel, AutoTokenizer
def process_infilling_prompt(prefix, suffix, tokenizer, number_of_mask):
prefix = [tokenizer.bos_token_id] + tokenizer.encode(prefix, add_special_tokens=False)
middle = [tokenizer.mask_token_id] * number_of_mask
suffix = tokenizer.encode(suffix, add_special_tokens=False) + [tokenizer.eos_token_id]
return prefix + middle + suffix
prefix = '''from typing import List
def has_close_elements(numbers: List[float], threshold: float) -> bool:
""" Check if in given list of numbers, are any two numbers closer to each other than
given threshold.
>>> has_close_elements([1.0, 2.0, 3.0], 0.5)
False
>>> has_close_elements([1.0, 2.8, 3.0, 4.0, 5.0, 2.0], 0.3)
True
"""
'''
suffix = ''' for idx2, elem2 in enumerate(numbers):
if idx != idx2:
distance = abs(elem - elem2)
if distance < threshold:
return True
return False
'''
model_path = 'Dream-org/DreamOn-v0-7B'
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
## Set the initial mask token length when processing the prompt
input_ids = process_infilling_prompt(prefix, suffix, tokenizer, 4)
input_ids = torch.LongTensor([input_ids]).to("cuda")
model = AutoModel.from_pretrained(model_path, torch_dtype=torch.bfloat16, trust_remote_code=True)
model = model.to("cuda").eval()
output = model.diffusion_generate(
input_ids,
temperature=0.2,
alg = 'entropy',
alg_temp = 0,
top_p = 0.9,
max_new_tokens = 64, ## Set the maximum number of new tokens for infilling
return_dict_in_generate = True,
output_history = True,
number_transfer_tokens = 1
)
history = output.history
for i, h in enumerate(history):
print(f"########################")
time.sleep(0.2)
print(tokenizer.decode(h.tolist()), end="\n\n")Quick Starttexttransformers
import torch
import time
from transformers import AutoModel, AutoTokenizer
def process_infilling_prompt(prefix, suffix, tokenizer, number_of_mask):
prefix = [tokenizer.bos_token_id] + tokenizer.encode(prefix, add_special_tokens=False)
middle = [tokenizer.mask_token_id] * number_of_mask
suffix = tokenizer.encode(suffix, add_special_tokens=False) + [tokenizer.eos_token_id]
return prefix + middle + suffix
prefix = '''from typing import List
def has_close_elements(numbers: List[float], threshold: float) -> bool:
""" Check if in given list of numbers, are any two numbers closer to each other than
given threshold.
>>> has_close_elements([1.0, 2.0, 3.0], 0.5)
False
>>> has_close_elements([1.0, 2.8, 3.0, 4.0, 5.0, 2.0], 0.3)
True
"""
'''
suffix = ''' for idx2, elem2 in enumerate(numbers):
if idx != idx2:
distance = abs(elem - elem2)
if distance < threshold:
return True
return False
'''
model_path = 'Dream-org/DreamOn-v0-7B'
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
## Set the initial mask token length when processing the prompt
input_ids = process_infilling_prompt(prefix, suffix, tokenizer, 4)
input_ids = torch.LongTensor([input_ids]).to("cuda")
model = AutoModel.from_pretrained(model_path, torch_dtype=torch.bfloat16, trust_remote_code=True)
model = model.to("cuda").eval()
output = model.diffusion_generate(
input_ids,
temperature=0.2,
alg = 'entropy',
alg_temp = 0,
top_p = 0.9,
max_new_tokens = 64, ## Set the maximum number of new tokens for infilling
return_dict_in_generate = True,
output_history = True,
number_transfer_tokens = 1
)
history = output.history
for i, h in enumerate(history):
print(f"########################")
time.sleep(0.2)
print(tokenizer.decode(h.tolist()), end="\n\n")Quick Starttexttransformers
import torch
import time
from transformers import AutoModel, AutoTokenizer
def process_infilling_prompt(prefix, suffix, tokenizer, number_of_mask):
prefix = [tokenizer.bos_token_id] + tokenizer.encode(prefix, add_special_tokens=False)
middle = [tokenizer.mask_token_id] * number_of_mask
suffix = tokenizer.encode(suffix, add_special_tokens=False) + [tokenizer.eos_token_id]
return prefix + middle + suffix
prefix = '''from typing import List
def has_close_elements(numbers: List[float], threshold: float) -> bool:
""" Check if in given list of numbers, are any two numbers closer to each other than
given threshold.
>>> has_close_elements([1.0, 2.0, 3.0], 0.5)
False
>>> has_close_elements([1.0, 2.8, 3.0, 4.0, 5.0, 2.0], 0.3)
True
"""
'''
suffix = ''' for idx2, elem2 in enumerate(numbers):
if idx != idx2:
distance = abs(elem - elem2)
if distance < threshold:
return True
return False
'''
model_path = 'Dream-org/DreamOn-v0-7B'
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
## Set the initial mask token length when processing the prompt
input_ids = process_infilling_prompt(prefix, suffix, tokenizer, 4)
input_ids = torch.LongTensor([input_ids]).to("cuda")
model = AutoModel.from_pretrained(model_path, torch_dtype=torch.bfloat16, trust_remote_code=True)
model = model.to("cuda").eval()
output = model.diffusion_generate(
input_ids,
temperature=0.2,
alg = 'entropy',
alg_temp = 0,
top_p = 0.9,
max_new_tokens = 64, ## Set the maximum number of new tokens for infilling
return_dict_in_generate = True,
output_history = True,
number_transfer_tokens = 1
)
history = output.history
for i, h in enumerate(history):
print(f"########################")
time.sleep(0.2)
print(tokenizer.decode(h.tolist()), end="\n\n")Quick Starttexttransformers
import torch
import time
from transformers import AutoModel, AutoTokenizer
def process_infilling_prompt(prefix, suffix, tokenizer, number_of_mask):
prefix = [tokenizer.bos_token_id] + tokenizer.encode(prefix, add_special_tokens=False)
middle = [tokenizer.mask_token_id] * number_of_mask
suffix = tokenizer.encode(suffix, add_special_tokens=False) + [tokenizer.eos_token_id]
return prefix + middle + suffix
prefix = '''from typing import List
def has_close_elements(numbers: List[float], threshold: float) -> bool:
""" Check if in given list of numbers, are any two numbers closer to each other than
given threshold.
>>> has_close_elements([1.0, 2.0, 3.0], 0.5)
False
>>> has_close_elements([1.0, 2.8, 3.0, 4.0, 5.0, 2.0], 0.3)
True
"""
'''
suffix = ''' for idx2, elem2 in enumerate(numbers):
if idx != idx2:
distance = abs(elem - elem2)
if distance < threshold:
return True
return False
'''
model_path = 'Dream-org/DreamOn-v0-7B'
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
## Set the initial mask token length when processing the prompt
input_ids = process_infilling_prompt(prefix, suffix, tokenizer, 4)
input_ids = torch.LongTensor([input_ids]).to("cuda")
model = AutoModel.from_pretrained(model_path, torch_dtype=torch.bfloat16, trust_remote_code=True)
model = model.to("cuda").eval()
output = model.diffusion_generate(
input_ids,
temperature=0.2,
alg = 'entropy',
alg_temp = 0,
top_p = 0.9,
max_new_tokens = 64, ## Set the maximum number of new tokens for infilling
return_dict_in_generate = True,
output_history = True,
number_transfer_tokens = 1
)
history = output.history
for i, h in enumerate(history):
print(f"########################")
time.sleep(0.2)
print(tokenizer.decode(h.tolist()), end="\n\n")Quick Starttexttransformers
import torch
import time
from transformers import AutoModel, AutoTokenizer
def process_infilling_prompt(prefix, suffix, tokenizer, number_of_mask):
prefix = [tokenizer.bos_token_id] + tokenizer.encode(prefix, add_special_tokens=False)
middle = [tokenizer.mask_token_id] * number_of_mask
suffix = tokenizer.encode(suffix, add_special_tokens=False) + [tokenizer.eos_token_id]
return prefix + middle + suffix
prefix = '''from typing import List
def has_close_elements(numbers: List[float], threshold: float) -> bool:
""" Check if in given list of numbers, are any two numbers closer to each other than
given threshold.
>>> has_close_elements([1.0, 2.0, 3.0], 0.5)
False
>>> has_close_elements([1.0, 2.8, 3.0, 4.0, 5.0, 2.0], 0.3)
True
"""
'''
suffix = ''' for idx2, elem2 in enumerate(numbers):
if idx != idx2:
distance = abs(elem - elem2)
if distance < threshold:
return True
return False
'''
model_path = 'Dream-org/DreamOn-v0-7B'
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
## Set the initial mask token length when processing the prompt
input_ids = process_infilling_prompt(prefix, suffix, tokenizer, 4)
input_ids = torch.LongTensor([input_ids]).to("cuda")
model = AutoModel.from_pretrained(model_path, torch_dtype=torch.bfloat16, trust_remote_code=True)
model = model.to("cuda").eval()
output = model.diffusion_generate(
input_ids,
temperature=0.2,
alg = 'entropy',
alg_temp = 0,
top_p = 0.9,
max_new_tokens = 64, ## Set the maximum number of new tokens for infilling
return_dict_in_generate = True,
output_history = True,
number_transfer_tokens = 1
)
history = output.history
for i, h in enumerate(history):
print(f"########################")
time.sleep(0.2)
print(tokenizer.decode(h.tolist()), end="\n\n")Quick Starttexttransformers
import torch
import time
from transformers import AutoModel, AutoTokenizer
def process_infilling_prompt(prefix, suffix, tokenizer, number_of_mask):
prefix = [tokenizer.bos_token_id] + tokenizer.encode(prefix, add_special_tokens=False)
middle = [tokenizer.mask_token_id] * number_of_mask
suffix = tokenizer.encode(suffix, add_special_tokens=False) + [tokenizer.eos_token_id]
return prefix + middle + suffix
prefix = '''from typing import List
def has_close_elements(numbers: List[float], threshold: float) -> bool:
""" Check if in given list of numbers, are any two numbers closer to each other than
given threshold.
>>> has_close_elements([1.0, 2.0, 3.0], 0.5)
False
>>> has_close_elements([1.0, 2.8, 3.0, 4.0, 5.0, 2.0], 0.3)
True
"""
'''
suffix = ''' for idx2, elem2 in enumerate(numbers):
if idx != idx2:
distance = abs(elem - elem2)
if distance < threshold:
return True
return False
'''
model_path = 'Dream-org/DreamOn-v0-7B'
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
## Set the initial mask token length when processing the prompt
input_ids = process_infilling_prompt(prefix, suffix, tokenizer, 4)
input_ids = torch.LongTensor([input_ids]).to("cuda")
model = AutoModel.from_pretrained(model_path, torch_dtype=torch.bfloat16, trust_remote_code=True)
model = model.to("cuda").eval()
output = model.diffusion_generate(
input_ids,
temperature=0.2,
alg = 'entropy',
alg_temp = 0,
top_p = 0.9,
max_new_tokens = 64, ## Set the maximum number of new tokens for infilling
return_dict_in_generate = True,
output_history = True,
number_transfer_tokens = 1
)
history = output.history
for i, h in enumerate(history):
print(f"########################")
time.sleep(0.2)
print(tokenizer.decode(h.tolist()), end="\n\n")Quick Starttexttransformers
import torch
import time
from transformers import AutoModel, AutoTokenizer
def process_infilling_prompt(prefix, suffix, tokenizer, number_of_mask):
prefix = [tokenizer.bos_token_id] + tokenizer.encode(prefix, add_special_tokens=False)
middle = [tokenizer.mask_token_id] * number_of_mask
suffix = tokenizer.encode(suffix, add_special_tokens=False) + [tokenizer.eos_token_id]
return prefix + middle + suffix
prefix = '''from typing import List
def has_close_elements(numbers: List[float], threshold: float) -> bool:
""" Check if in given list of numbers, are any two numbers closer to each other than
given threshold.
>>> has_close_elements([1.0, 2.0, 3.0], 0.5)
False
>>> has_close_elements([1.0, 2.8, 3.0, 4.0, 5.0, 2.0], 0.3)
True
"""
'''
suffix = ''' for idx2, elem2 in enumerate(numbers):
if idx != idx2:
distance = abs(elem - elem2)
if distance < threshold:
return True
return False
'''
model_path = 'Dream-org/DreamOn-v0-7B'
tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=True)
## Set the initial mask token length when processing the prompt
input_ids = process_infilling_prompt(prefix, suffix, tokenizer, 4)
input_ids = torch.LongTensor([input_ids]).to("cuda")
model = AutoModel.from_pretrained(model_path, torch_dtype=torch.bfloat16, trust_remote_code=True)
model = model.to("cuda").eval()
output = model.diffusion_generate(
input_ids,
temperature=0.2,
alg = 'entropy',
alg_temp = 0,
top_p = 0.9,
max_new_tokens = 64, ## Set the maximum number of new tokens for infilling
return_dict_in_generate = True,
output_history = True,
number_transfer_tokens = 1
)
history = output.history
for i, h in enumerate(history):
print(f"########################")
time.sleep(0.2)
print(tokenizer.decode(h.tolist()), end="\n\n")Deploy This Model
Production-ready deployment in minutes
Together.ai
Instant API access to this model
Production-ready inference API. Start free, scale to millions.
Try Free APIReplicate
One-click model deployment
Run models in the cloud with simple API. No DevOps required.
Deploy NowDisclosure: We may earn a commission from these partners. This helps keep LLMYourWay free.