yizhao-risk-en-scorer

4
license:apache-2.0
by
HIT-TMG
Other
OTHER
New
0 downloads
Early-stage
Edge AI:
Mobile
Laptop
Server
Unknown
Mobile
Laptop
Server
Quick Summary

yizhao-risk-en-scorer Introduction This is a BERT model fine-tuned on a high-quality English financial dataset.

Code Examples

Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}
Quickstartpythontransformers
from transformers import AutoTokenizer, AutoModelForSequenceClassification

text = "You are a smart robot"
risk_model_name = "risk-model-en-v0.1"

risk_tokenizer = AutoTokenizer.from_pretrained(risk_model_name)
risk_model = AutoModelForSequenceClassification.from_pretrained(risk_model_name)

risk_inputs = risk_tokenizer(text, return_tensors="pt", padding="longest", truncation=True)
risk_outputs = risk_model(**risk_inputs)
risk_logits = risk_outputs.logits.squeeze(-1).float().detach().numpy()

risk_score = risk_logits.item()

result = {
    "text": text,
    "risk_score": risk_score
}

print(result)
# {'text': 'You are a smart robot', 'risk_score': 0.11226219683885574}

Deploy This Model

Production-ready deployment in minutes

Together.ai

Instant API access to this model

Fastest API

Production-ready inference API. Start free, scale to millions.

Try Free API

Replicate

One-click model deployment

Easiest Setup

Run models in the cloud with simple API. No DevOps required.

Deploy Now

Disclosure: We may earn a commission from these partners. This helps keep LLMYourWay free.