textdetox

11 models • 3 total models in database
Sort by:

xlmr-large-toxicity-classifier-v2

23,823
9

xlmr-large-toxicity-classifier

Multilingual Toxicity Classifier for 9 languages (2024) This is an instance of xlm-roberta-large that was fine-tuned on binary toxicity classification task based on our compiled dataset textdetox/multilingualtoxicitydataset. Firstly, we separated a balanced 20% test set to check the model adequency. Then, the model was fine-tuned on the full data. The results on the test set are the following: | | Precision | Recall | F1 | |----------|-----------|--------|-------| | alllang | 0.8713 | 0.8710 | 0.8710| | en | 0.9650 | 0.9650 | 0.9650| | ru | 0.9791 | 0.9790 | 0.9790| | uk | 0.9267 | 0.9250 | 0.9251| | de | 0.8791 | 0.8760 | 0.8758| | es | 0.8700 | 0.8700 | 0.8700| | ar | 0.7787 | 0.7780 | 0.7780| | am | 0.7781 | 0.7780 | 0.7780| | hi | 0.9360 | 0.9360 | 0.9360| | zh | 0.7318 | 0.7320 | 0.7315| Citation If you would like to acknowledge our work, please, cite the following manuscripts:

1,851
15

bert-multilingual-toxicity-classifier

1,408
5

Glot500 Toxicity Classifier

206
1

twitter-xlmr-toxicity-classifier

NaNK
111
0

xlmr-base-toxicity-classifier

31
2

Llama-pairwise-content-evaluator

This model is a fine-tuned version of meta-llama/Llama-3.1-8B-Instruct. It has been trained using TRL. - TRL: 0.16.0 - Transformers: 4.50.1 - Pytorch: 2.5.1 - Datasets: 3.4.1 - Tokenizers: 0.21.1

NaNK
llama
11
0

Llama-pairwise-toxicity-evaluator

This model is a fine-tuned version of meta-llama/Llama-3.1-8B-Instruct. It has been trained using TRL with textdetox/detoxificationpairwisestyleevaluation dataset - TRL: 0.16.0 - Transformers: 4.50.1 - Pytorch: 2.5.1 - Datasets: 3.4.1 - Tokenizers: 0.21.1

NaNK
llama
5
0

mbart-detox-baseline

NaNK
2
0

mbart_detox_en_ru_uk_es

NaNK
1
2

mt5-xl-detox-baseline

license:mit
0
2