eai-distill-0.5b-GGUF
428
2
500M
Q4
license:apache-2.0
by
QuantFactory
Other
OTHER
0.5B params
New
428 downloads
Early-stage
Edge AI:
Mobile
Laptop
Server
2GB+ RAM
Mobile
Laptop
Server
Quick Summary
QuantFactory/eai-distill-0.
Device Compatibility
Mobile
4-6GB RAM
Laptop
16GB RAM
Server
GPU
Minimum Recommended
1GB+ RAM
Code Examples
đ» Usagepythontransformers
from transformers import AutoTokenizer, AutoModelForCausalLM
import random
# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("EssentialAI/EAI-Distill-0.5b", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("EssentialAI/EAI-Distill-0.5b")
def chunk_text(text, max_char_per_doc=30000):
if len(text) <= max_char_per_doc:
return text
chunk_size = max_char_per_doc // 3
start = text[:chunk_size]
middle_start = chunk_size
middle_end = len(text) - chunk_size
mid_point = random.randint(middle_start + chunk_size//2, middle_end - chunk_size//2)
middle = text[mid_point - chunk_size//2:mid_point + chunk_size//2]
end = text[-chunk_size:]
return f"[beginning]\n{start}\n[middle]\n{middle}\n[end]\n{end}"
def classify_document(text):
chunked_text = chunk_text(text)
messages = [
{"role": "system", "content": "taxonomy"},
{"role": "user", "content": chunked_text},
]
prompt = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Example usage
document_text = "Your document content here..."
classification = classify_document(document_text)
print(classification)đ» Usagepythontransformers
from transformers import AutoTokenizer, AutoModelForCausalLM
import random
# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("EssentialAI/EAI-Distill-0.5b", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("EssentialAI/EAI-Distill-0.5b")
def chunk_text(text, max_char_per_doc=30000):
if len(text) <= max_char_per_doc:
return text
chunk_size = max_char_per_doc // 3
start = text[:chunk_size]
middle_start = chunk_size
middle_end = len(text) - chunk_size
mid_point = random.randint(middle_start + chunk_size//2, middle_end - chunk_size//2)
middle = text[mid_point - chunk_size//2:mid_point + chunk_size//2]
end = text[-chunk_size:]
return f"[beginning]\n{start}\n[middle]\n{middle}\n[end]\n{end}"
def classify_document(text):
chunked_text = chunk_text(text)
messages = [
{"role": "system", "content": "taxonomy"},
{"role": "user", "content": chunked_text},
]
prompt = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Example usage
document_text = "Your document content here..."
classification = classify_document(document_text)
print(classification)đ» Usagepythontransformers
from transformers import AutoTokenizer, AutoModelForCausalLM
import random
# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("EssentialAI/EAI-Distill-0.5b", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("EssentialAI/EAI-Distill-0.5b")
def chunk_text(text, max_char_per_doc=30000):
if len(text) <= max_char_per_doc:
return text
chunk_size = max_char_per_doc // 3
start = text[:chunk_size]
middle_start = chunk_size
middle_end = len(text) - chunk_size
mid_point = random.randint(middle_start + chunk_size//2, middle_end - chunk_size//2)
middle = text[mid_point - chunk_size//2:mid_point + chunk_size//2]
end = text[-chunk_size:]
return f"[beginning]\n{start}\n[middle]\n{middle}\n[end]\n{end}"
def classify_document(text):
chunked_text = chunk_text(text)
messages = [
{"role": "system", "content": "taxonomy"},
{"role": "user", "content": chunked_text},
]
prompt = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Example usage
document_text = "Your document content here..."
classification = classify_document(document_text)
print(classification)đ» Usagepythontransformers
from transformers import AutoTokenizer, AutoModelForCausalLM
import random
# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("EssentialAI/EAI-Distill-0.5b", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("EssentialAI/EAI-Distill-0.5b")
def chunk_text(text, max_char_per_doc=30000):
if len(text) <= max_char_per_doc:
return text
chunk_size = max_char_per_doc // 3
start = text[:chunk_size]
middle_start = chunk_size
middle_end = len(text) - chunk_size
mid_point = random.randint(middle_start + chunk_size//2, middle_end - chunk_size//2)
middle = text[mid_point - chunk_size//2:mid_point + chunk_size//2]
end = text[-chunk_size:]
return f"[beginning]\n{start}\n[middle]\n{middle}\n[end]\n{end}"
def classify_document(text):
chunked_text = chunk_text(text)
messages = [
{"role": "system", "content": "taxonomy"},
{"role": "user", "content": chunked_text},
]
prompt = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Example usage
document_text = "Your document content here..."
classification = classify_document(document_text)
print(classification)đ» Usagepythontransformers
from transformers import AutoTokenizer, AutoModelForCausalLM
import random
# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("EssentialAI/EAI-Distill-0.5b", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("EssentialAI/EAI-Distill-0.5b")
def chunk_text(text, max_char_per_doc=30000):
if len(text) <= max_char_per_doc:
return text
chunk_size = max_char_per_doc // 3
start = text[:chunk_size]
middle_start = chunk_size
middle_end = len(text) - chunk_size
mid_point = random.randint(middle_start + chunk_size//2, middle_end - chunk_size//2)
middle = text[mid_point - chunk_size//2:mid_point + chunk_size//2]
end = text[-chunk_size:]
return f"[beginning]\n{start}\n[middle]\n{middle}\n[end]\n{end}"
def classify_document(text):
chunked_text = chunk_text(text)
messages = [
{"role": "system", "content": "taxonomy"},
{"role": "user", "content": chunked_text},
]
prompt = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Example usage
document_text = "Your document content here..."
classification = classify_document(document_text)
print(classification)đ» Usagepythontransformers
from transformers import AutoTokenizer, AutoModelForCausalLM
import random
# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("EssentialAI/EAI-Distill-0.5b", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("EssentialAI/EAI-Distill-0.5b")
def chunk_text(text, max_char_per_doc=30000):
if len(text) <= max_char_per_doc:
return text
chunk_size = max_char_per_doc // 3
start = text[:chunk_size]
middle_start = chunk_size
middle_end = len(text) - chunk_size
mid_point = random.randint(middle_start + chunk_size//2, middle_end - chunk_size//2)
middle = text[mid_point - chunk_size//2:mid_point + chunk_size//2]
end = text[-chunk_size:]
return f"[beginning]\n{start}\n[middle]\n{middle}\n[end]\n{end}"
def classify_document(text):
chunked_text = chunk_text(text)
messages = [
{"role": "system", "content": "taxonomy"},
{"role": "user", "content": chunked_text},
]
prompt = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Example usage
document_text = "Your document content here..."
classification = classify_document(document_text)
print(classification)đ» Usagepythontransformers
from transformers import AutoTokenizer, AutoModelForCausalLM
import random
# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("EssentialAI/EAI-Distill-0.5b", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("EssentialAI/EAI-Distill-0.5b")
def chunk_text(text, max_char_per_doc=30000):
if len(text) <= max_char_per_doc:
return text
chunk_size = max_char_per_doc // 3
start = text[:chunk_size]
middle_start = chunk_size
middle_end = len(text) - chunk_size
mid_point = random.randint(middle_start + chunk_size//2, middle_end - chunk_size//2)
middle = text[mid_point - chunk_size//2:mid_point + chunk_size//2]
end = text[-chunk_size:]
return f"[beginning]\n{start}\n[middle]\n{middle}\n[end]\n{end}"
def classify_document(text):
chunked_text = chunk_text(text)
messages = [
{"role": "system", "content": "taxonomy"},
{"role": "user", "content": chunked_text},
]
prompt = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Example usage
document_text = "Your document content here..."
classification = classify_document(document_text)
print(classification)đ» Usagepythontransformers
from transformers import AutoTokenizer, AutoModelForCausalLM
import random
# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("EssentialAI/EAI-Distill-0.5b", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("EssentialAI/EAI-Distill-0.5b")
def chunk_text(text, max_char_per_doc=30000):
if len(text) <= max_char_per_doc:
return text
chunk_size = max_char_per_doc // 3
start = text[:chunk_size]
middle_start = chunk_size
middle_end = len(text) - chunk_size
mid_point = random.randint(middle_start + chunk_size//2, middle_end - chunk_size//2)
middle = text[mid_point - chunk_size//2:mid_point + chunk_size//2]
end = text[-chunk_size:]
return f"[beginning]\n{start}\n[middle]\n{middle}\n[end]\n{end}"
def classify_document(text):
chunked_text = chunk_text(text)
messages = [
{"role": "system", "content": "taxonomy"},
{"role": "user", "content": chunked_text},
]
prompt = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Example usage
document_text = "Your document content here..."
classification = classify_document(document_text)
print(classification)đ» Usagepythontransformers
from transformers import AutoTokenizer, AutoModelForCausalLM
import random
# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("EssentialAI/EAI-Distill-0.5b", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("EssentialAI/EAI-Distill-0.5b")
def chunk_text(text, max_char_per_doc=30000):
if len(text) <= max_char_per_doc:
return text
chunk_size = max_char_per_doc // 3
start = text[:chunk_size]
middle_start = chunk_size
middle_end = len(text) - chunk_size
mid_point = random.randint(middle_start + chunk_size//2, middle_end - chunk_size//2)
middle = text[mid_point - chunk_size//2:mid_point + chunk_size//2]
end = text[-chunk_size:]
return f"[beginning]\n{start}\n[middle]\n{middle}\n[end]\n{end}"
def classify_document(text):
chunked_text = chunk_text(text)
messages = [
{"role": "system", "content": "taxonomy"},
{"role": "user", "content": chunked_text},
]
prompt = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Example usage
document_text = "Your document content here..."
classification = classify_document(document_text)
print(classification)đ» Usagepythontransformers
from transformers import AutoTokenizer, AutoModelForCausalLM
import random
# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("EssentialAI/EAI-Distill-0.5b", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("EssentialAI/EAI-Distill-0.5b")
def chunk_text(text, max_char_per_doc=30000):
if len(text) <= max_char_per_doc:
return text
chunk_size = max_char_per_doc // 3
start = text[:chunk_size]
middle_start = chunk_size
middle_end = len(text) - chunk_size
mid_point = random.randint(middle_start + chunk_size//2, middle_end - chunk_size//2)
middle = text[mid_point - chunk_size//2:mid_point + chunk_size//2]
end = text[-chunk_size:]
return f"[beginning]\n{start}\n[middle]\n{middle}\n[end]\n{end}"
def classify_document(text):
chunked_text = chunk_text(text)
messages = [
{"role": "system", "content": "taxonomy"},
{"role": "user", "content": chunked_text},
]
prompt = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Example usage
document_text = "Your document content here..."
classification = classify_document(document_text)
print(classification)đ» Usagepythontransformers
from transformers import AutoTokenizer, AutoModelForCausalLM
import random
# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("EssentialAI/EAI-Distill-0.5b", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("EssentialAI/EAI-Distill-0.5b")
def chunk_text(text, max_char_per_doc=30000):
if len(text) <= max_char_per_doc:
return text
chunk_size = max_char_per_doc // 3
start = text[:chunk_size]
middle_start = chunk_size
middle_end = len(text) - chunk_size
mid_point = random.randint(middle_start + chunk_size//2, middle_end - chunk_size//2)
middle = text[mid_point - chunk_size//2:mid_point + chunk_size//2]
end = text[-chunk_size:]
return f"[beginning]\n{start}\n[middle]\n{middle}\n[end]\n{end}"
def classify_document(text):
chunked_text = chunk_text(text)
messages = [
{"role": "system", "content": "taxonomy"},
{"role": "user", "content": chunked_text},
]
prompt = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Example usage
document_text = "Your document content here..."
classification = classify_document(document_text)
print(classification)đ» Usagepythontransformers
from transformers import AutoTokenizer, AutoModelForCausalLM
import random
# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("EssentialAI/EAI-Distill-0.5b", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("EssentialAI/EAI-Distill-0.5b")
def chunk_text(text, max_char_per_doc=30000):
if len(text) <= max_char_per_doc:
return text
chunk_size = max_char_per_doc // 3
start = text[:chunk_size]
middle_start = chunk_size
middle_end = len(text) - chunk_size
mid_point = random.randint(middle_start + chunk_size//2, middle_end - chunk_size//2)
middle = text[mid_point - chunk_size//2:mid_point + chunk_size//2]
end = text[-chunk_size:]
return f"[beginning]\n{start}\n[middle]\n{middle}\n[end]\n{end}"
def classify_document(text):
chunked_text = chunk_text(text)
messages = [
{"role": "system", "content": "taxonomy"},
{"role": "user", "content": chunked_text},
]
prompt = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Example usage
document_text = "Your document content here..."
classification = classify_document(document_text)
print(classification)đ» Usagepythontransformers
from transformers import AutoTokenizer, AutoModelForCausalLM
import random
# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("EssentialAI/EAI-Distill-0.5b", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("EssentialAI/EAI-Distill-0.5b")
def chunk_text(text, max_char_per_doc=30000):
if len(text) <= max_char_per_doc:
return text
chunk_size = max_char_per_doc // 3
start = text[:chunk_size]
middle_start = chunk_size
middle_end = len(text) - chunk_size
mid_point = random.randint(middle_start + chunk_size//2, middle_end - chunk_size//2)
middle = text[mid_point - chunk_size//2:mid_point + chunk_size//2]
end = text[-chunk_size:]
return f"[beginning]\n{start}\n[middle]\n{middle}\n[end]\n{end}"
def classify_document(text):
chunked_text = chunk_text(text)
messages = [
{"role": "system", "content": "taxonomy"},
{"role": "user", "content": chunked_text},
]
prompt = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Example usage
document_text = "Your document content here..."
classification = classify_document(document_text)
print(classification)đ» Usagepythontransformers
from transformers import AutoTokenizer, AutoModelForCausalLM
import random
# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("EssentialAI/EAI-Distill-0.5b", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("EssentialAI/EAI-Distill-0.5b")
def chunk_text(text, max_char_per_doc=30000):
if len(text) <= max_char_per_doc:
return text
chunk_size = max_char_per_doc // 3
start = text[:chunk_size]
middle_start = chunk_size
middle_end = len(text) - chunk_size
mid_point = random.randint(middle_start + chunk_size//2, middle_end - chunk_size//2)
middle = text[mid_point - chunk_size//2:mid_point + chunk_size//2]
end = text[-chunk_size:]
return f"[beginning]\n{start}\n[middle]\n{middle}\n[end]\n{end}"
def classify_document(text):
chunked_text = chunk_text(text)
messages = [
{"role": "system", "content": "taxonomy"},
{"role": "user", "content": chunked_text},
]
prompt = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Example usage
document_text = "Your document content here..."
classification = classify_document(document_text)
print(classification)đ» Usagepythontransformers
from transformers import AutoTokenizer, AutoModelForCausalLM
import random
# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("EssentialAI/EAI-Distill-0.5b", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("EssentialAI/EAI-Distill-0.5b")
def chunk_text(text, max_char_per_doc=30000):
if len(text) <= max_char_per_doc:
return text
chunk_size = max_char_per_doc // 3
start = text[:chunk_size]
middle_start = chunk_size
middle_end = len(text) - chunk_size
mid_point = random.randint(middle_start + chunk_size//2, middle_end - chunk_size//2)
middle = text[mid_point - chunk_size//2:mid_point + chunk_size//2]
end = text[-chunk_size:]
return f"[beginning]\n{start}\n[middle]\n{middle}\n[end]\n{end}"
def classify_document(text):
chunked_text = chunk_text(text)
messages = [
{"role": "system", "content": "taxonomy"},
{"role": "user", "content": chunked_text},
]
prompt = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Example usage
document_text = "Your document content here..."
classification = classify_document(document_text)
print(classification)đ» Usagepythontransformers
from transformers import AutoTokenizer, AutoModelForCausalLM
import random
# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("EssentialAI/EAI-Distill-0.5b", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("EssentialAI/EAI-Distill-0.5b")
def chunk_text(text, max_char_per_doc=30000):
if len(text) <= max_char_per_doc:
return text
chunk_size = max_char_per_doc // 3
start = text[:chunk_size]
middle_start = chunk_size
middle_end = len(text) - chunk_size
mid_point = random.randint(middle_start + chunk_size//2, middle_end - chunk_size//2)
middle = text[mid_point - chunk_size//2:mid_point + chunk_size//2]
end = text[-chunk_size:]
return f"[beginning]\n{start}\n[middle]\n{middle}\n[end]\n{end}"
def classify_document(text):
chunked_text = chunk_text(text)
messages = [
{"role": "system", "content": "taxonomy"},
{"role": "user", "content": chunked_text},
]
prompt = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Example usage
document_text = "Your document content here..."
classification = classify_document(document_text)
print(classification)đ» Usagepythontransformers
from transformers import AutoTokenizer, AutoModelForCausalLM
import random
# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("EssentialAI/EAI-Distill-0.5b", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("EssentialAI/EAI-Distill-0.5b")
def chunk_text(text, max_char_per_doc=30000):
if len(text) <= max_char_per_doc:
return text
chunk_size = max_char_per_doc // 3
start = text[:chunk_size]
middle_start = chunk_size
middle_end = len(text) - chunk_size
mid_point = random.randint(middle_start + chunk_size//2, middle_end - chunk_size//2)
middle = text[mid_point - chunk_size//2:mid_point + chunk_size//2]
end = text[-chunk_size:]
return f"[beginning]\n{start}\n[middle]\n{middle}\n[end]\n{end}"
def classify_document(text):
chunked_text = chunk_text(text)
messages = [
{"role": "system", "content": "taxonomy"},
{"role": "user", "content": chunked_text},
]
prompt = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Example usage
document_text = "Your document content here..."
classification = classify_document(document_text)
print(classification)đ» Usagepythontransformers
from transformers import AutoTokenizer, AutoModelForCausalLM
import random
# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("EssentialAI/EAI-Distill-0.5b", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("EssentialAI/EAI-Distill-0.5b")
def chunk_text(text, max_char_per_doc=30000):
if len(text) <= max_char_per_doc:
return text
chunk_size = max_char_per_doc // 3
start = text[:chunk_size]
middle_start = chunk_size
middle_end = len(text) - chunk_size
mid_point = random.randint(middle_start + chunk_size//2, middle_end - chunk_size//2)
middle = text[mid_point - chunk_size//2:mid_point + chunk_size//2]
end = text[-chunk_size:]
return f"[beginning]\n{start}\n[middle]\n{middle}\n[end]\n{end}"
def classify_document(text):
chunked_text = chunk_text(text)
messages = [
{"role": "system", "content": "taxonomy"},
{"role": "user", "content": chunked_text},
]
prompt = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Example usage
document_text = "Your document content here..."
classification = classify_document(document_text)
print(classification)đ» Usagepythontransformers
from transformers import AutoTokenizer, AutoModelForCausalLM
import random
# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("EssentialAI/EAI-Distill-0.5b", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("EssentialAI/EAI-Distill-0.5b")
def chunk_text(text, max_char_per_doc=30000):
if len(text) <= max_char_per_doc:
return text
chunk_size = max_char_per_doc // 3
start = text[:chunk_size]
middle_start = chunk_size
middle_end = len(text) - chunk_size
mid_point = random.randint(middle_start + chunk_size//2, middle_end - chunk_size//2)
middle = text[mid_point - chunk_size//2:mid_point + chunk_size//2]
end = text[-chunk_size:]
return f"[beginning]\n{start}\n[middle]\n{middle}\n[end]\n{end}"
def classify_document(text):
chunked_text = chunk_text(text)
messages = [
{"role": "system", "content": "taxonomy"},
{"role": "user", "content": chunked_text},
]
prompt = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Example usage
document_text = "Your document content here..."
classification = classify_document(document_text)
print(classification)đ» Usagepythontransformers
from transformers import AutoTokenizer, AutoModelForCausalLM
import random
# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("EssentialAI/EAI-Distill-0.5b", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("EssentialAI/EAI-Distill-0.5b")
def chunk_text(text, max_char_per_doc=30000):
if len(text) <= max_char_per_doc:
return text
chunk_size = max_char_per_doc // 3
start = text[:chunk_size]
middle_start = chunk_size
middle_end = len(text) - chunk_size
mid_point = random.randint(middle_start + chunk_size//2, middle_end - chunk_size//2)
middle = text[mid_point - chunk_size//2:mid_point + chunk_size//2]
end = text[-chunk_size:]
return f"[beginning]\n{start}\n[middle]\n{middle}\n[end]\n{end}"
def classify_document(text):
chunked_text = chunk_text(text)
messages = [
{"role": "system", "content": "taxonomy"},
{"role": "user", "content": chunked_text},
]
prompt = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Example usage
document_text = "Your document content here..."
classification = classify_document(document_text)
print(classification)đ» Usagepythontransformers
from transformers import AutoTokenizer, AutoModelForCausalLM
import random
# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("EssentialAI/EAI-Distill-0.5b", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("EssentialAI/EAI-Distill-0.5b")
def chunk_text(text, max_char_per_doc=30000):
if len(text) <= max_char_per_doc:
return text
chunk_size = max_char_per_doc // 3
start = text[:chunk_size]
middle_start = chunk_size
middle_end = len(text) - chunk_size
mid_point = random.randint(middle_start + chunk_size//2, middle_end - chunk_size//2)
middle = text[mid_point - chunk_size//2:mid_point + chunk_size//2]
end = text[-chunk_size:]
return f"[beginning]\n{start}\n[middle]\n{middle}\n[end]\n{end}"
def classify_document(text):
chunked_text = chunk_text(text)
messages = [
{"role": "system", "content": "taxonomy"},
{"role": "user", "content": chunked_text},
]
prompt = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Example usage
document_text = "Your document content here..."
classification = classify_document(document_text)
print(classification)đ» Usagepythontransformers
from transformers import AutoTokenizer, AutoModelForCausalLM
import random
# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("EssentialAI/EAI-Distill-0.5b", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("EssentialAI/EAI-Distill-0.5b")
def chunk_text(text, max_char_per_doc=30000):
if len(text) <= max_char_per_doc:
return text
chunk_size = max_char_per_doc // 3
start = text[:chunk_size]
middle_start = chunk_size
middle_end = len(text) - chunk_size
mid_point = random.randint(middle_start + chunk_size//2, middle_end - chunk_size//2)
middle = text[mid_point - chunk_size//2:mid_point + chunk_size//2]
end = text[-chunk_size:]
return f"[beginning]\n{start}\n[middle]\n{middle}\n[end]\n{end}"
def classify_document(text):
chunked_text = chunk_text(text)
messages = [
{"role": "system", "content": "taxonomy"},
{"role": "user", "content": chunked_text},
]
prompt = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Example usage
document_text = "Your document content here..."
classification = classify_document(document_text)
print(classification)đ» Usagepythontransformers
from transformers import AutoTokenizer, AutoModelForCausalLM
import random
# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("EssentialAI/EAI-Distill-0.5b", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("EssentialAI/EAI-Distill-0.5b")
def chunk_text(text, max_char_per_doc=30000):
if len(text) <= max_char_per_doc:
return text
chunk_size = max_char_per_doc // 3
start = text[:chunk_size]
middle_start = chunk_size
middle_end = len(text) - chunk_size
mid_point = random.randint(middle_start + chunk_size//2, middle_end - chunk_size//2)
middle = text[mid_point - chunk_size//2:mid_point + chunk_size//2]
end = text[-chunk_size:]
return f"[beginning]\n{start}\n[middle]\n{middle}\n[end]\n{end}"
def classify_document(text):
chunked_text = chunk_text(text)
messages = [
{"role": "system", "content": "taxonomy"},
{"role": "user", "content": chunked_text},
]
prompt = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Example usage
document_text = "Your document content here..."
classification = classify_document(document_text)
print(classification)đ» Usagepythontransformers
from transformers import AutoTokenizer, AutoModelForCausalLM
import random
# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("EssentialAI/EAI-Distill-0.5b", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("EssentialAI/EAI-Distill-0.5b")
def chunk_text(text, max_char_per_doc=30000):
if len(text) <= max_char_per_doc:
return text
chunk_size = max_char_per_doc // 3
start = text[:chunk_size]
middle_start = chunk_size
middle_end = len(text) - chunk_size
mid_point = random.randint(middle_start + chunk_size//2, middle_end - chunk_size//2)
middle = text[mid_point - chunk_size//2:mid_point + chunk_size//2]
end = text[-chunk_size:]
return f"[beginning]\n{start}\n[middle]\n{middle}\n[end]\n{end}"
def classify_document(text):
chunked_text = chunk_text(text)
messages = [
{"role": "system", "content": "taxonomy"},
{"role": "user", "content": chunked_text},
]
prompt = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Example usage
document_text = "Your document content here..."
classification = classify_document(document_text)
print(classification)đ» Usagepythontransformers
from transformers import AutoTokenizer, AutoModelForCausalLM
import random
# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("EssentialAI/EAI-Distill-0.5b", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("EssentialAI/EAI-Distill-0.5b")
def chunk_text(text, max_char_per_doc=30000):
if len(text) <= max_char_per_doc:
return text
chunk_size = max_char_per_doc // 3
start = text[:chunk_size]
middle_start = chunk_size
middle_end = len(text) - chunk_size
mid_point = random.randint(middle_start + chunk_size//2, middle_end - chunk_size//2)
middle = text[mid_point - chunk_size//2:mid_point + chunk_size//2]
end = text[-chunk_size:]
return f"[beginning]\n{start}\n[middle]\n{middle}\n[end]\n{end}"
def classify_document(text):
chunked_text = chunk_text(text)
messages = [
{"role": "system", "content": "taxonomy"},
{"role": "user", "content": chunked_text},
]
prompt = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Example usage
document_text = "Your document content here..."
classification = classify_document(document_text)
print(classification)đ» Usagepythontransformers
from transformers import AutoTokenizer, AutoModelForCausalLM
import random
# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("EssentialAI/EAI-Distill-0.5b", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("EssentialAI/EAI-Distill-0.5b")
def chunk_text(text, max_char_per_doc=30000):
if len(text) <= max_char_per_doc:
return text
chunk_size = max_char_per_doc // 3
start = text[:chunk_size]
middle_start = chunk_size
middle_end = len(text) - chunk_size
mid_point = random.randint(middle_start + chunk_size//2, middle_end - chunk_size//2)
middle = text[mid_point - chunk_size//2:mid_point + chunk_size//2]
end = text[-chunk_size:]
return f"[beginning]\n{start}\n[middle]\n{middle}\n[end]\n{end}"
def classify_document(text):
chunked_text = chunk_text(text)
messages = [
{"role": "system", "content": "taxonomy"},
{"role": "user", "content": chunked_text},
]
prompt = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Example usage
document_text = "Your document content here..."
classification = classify_document(document_text)
print(classification)đ» Usagepythontransformers
from transformers import AutoTokenizer, AutoModelForCausalLM
import random
# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("EssentialAI/EAI-Distill-0.5b", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("EssentialAI/EAI-Distill-0.5b")
def chunk_text(text, max_char_per_doc=30000):
if len(text) <= max_char_per_doc:
return text
chunk_size = max_char_per_doc // 3
start = text[:chunk_size]
middle_start = chunk_size
middle_end = len(text) - chunk_size
mid_point = random.randint(middle_start + chunk_size//2, middle_end - chunk_size//2)
middle = text[mid_point - chunk_size//2:mid_point + chunk_size//2]
end = text[-chunk_size:]
return f"[beginning]\n{start}\n[middle]\n{middle}\n[end]\n{end}"
def classify_document(text):
chunked_text = chunk_text(text)
messages = [
{"role": "system", "content": "taxonomy"},
{"role": "user", "content": chunked_text},
]
prompt = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Example usage
document_text = "Your document content here..."
classification = classify_document(document_text)
print(classification)đ» Usagepythontransformers
from transformers import AutoTokenizer, AutoModelForCausalLM
import random
# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("EssentialAI/EAI-Distill-0.5b", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("EssentialAI/EAI-Distill-0.5b")
def chunk_text(text, max_char_per_doc=30000):
if len(text) <= max_char_per_doc:
return text
chunk_size = max_char_per_doc // 3
start = text[:chunk_size]
middle_start = chunk_size
middle_end = len(text) - chunk_size
mid_point = random.randint(middle_start + chunk_size//2, middle_end - chunk_size//2)
middle = text[mid_point - chunk_size//2:mid_point + chunk_size//2]
end = text[-chunk_size:]
return f"[beginning]\n{start}\n[middle]\n{middle}\n[end]\n{end}"
def classify_document(text):
chunked_text = chunk_text(text)
messages = [
{"role": "system", "content": "taxonomy"},
{"role": "user", "content": chunked_text},
]
prompt = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Example usage
document_text = "Your document content here..."
classification = classify_document(document_text)
print(classification)đ» Usagepythontransformers
from transformers import AutoTokenizer, AutoModelForCausalLM
import random
# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("EssentialAI/EAI-Distill-0.5b", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("EssentialAI/EAI-Distill-0.5b")
def chunk_text(text, max_char_per_doc=30000):
if len(text) <= max_char_per_doc:
return text
chunk_size = max_char_per_doc // 3
start = text[:chunk_size]
middle_start = chunk_size
middle_end = len(text) - chunk_size
mid_point = random.randint(middle_start + chunk_size//2, middle_end - chunk_size//2)
middle = text[mid_point - chunk_size//2:mid_point + chunk_size//2]
end = text[-chunk_size:]
return f"[beginning]\n{start}\n[middle]\n{middle}\n[end]\n{end}"
def classify_document(text):
chunked_text = chunk_text(text)
messages = [
{"role": "system", "content": "taxonomy"},
{"role": "user", "content": chunked_text},
]
prompt = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Example usage
document_text = "Your document content here..."
classification = classify_document(document_text)
print(classification)đ» Usagepythontransformers
from transformers import AutoTokenizer, AutoModelForCausalLM
import random
# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("EssentialAI/EAI-Distill-0.5b", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("EssentialAI/EAI-Distill-0.5b")
def chunk_text(text, max_char_per_doc=30000):
if len(text) <= max_char_per_doc:
return text
chunk_size = max_char_per_doc // 3
start = text[:chunk_size]
middle_start = chunk_size
middle_end = len(text) - chunk_size
mid_point = random.randint(middle_start + chunk_size//2, middle_end - chunk_size//2)
middle = text[mid_point - chunk_size//2:mid_point + chunk_size//2]
end = text[-chunk_size:]
return f"[beginning]\n{start}\n[middle]\n{middle}\n[end]\n{end}"
def classify_document(text):
chunked_text = chunk_text(text)
messages = [
{"role": "system", "content": "taxonomy"},
{"role": "user", "content": chunked_text},
]
prompt = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
return tokenizer.decode(outputs[0], skip_special_tokens=True)
# Example usage
document_text = "Your document content here..."
classification = classify_document(document_text)
print(classification)Deploy This Model
Production-ready deployment in minutes
Together.ai
Instant API access to this model
Production-ready inference API. Start free, scale to millions.
Try Free APIReplicate
One-click model deployment
Run models in the cloud with simple API. No DevOps required.
Deploy NowDisclosure: We may earn a commission from these partners. This helps keep LLMYourWay free.