textdetox
xlmr-large-toxicity-classifier-v2
xlmr-large-toxicity-classifier
Multilingual Toxicity Classifier for 9 languages (2024) This is an instance of xlm-roberta-large that was fine-tuned on binary toxicity classification task based on our compiled dataset textdetox/multilingualtoxicitydataset. Firstly, we separated a balanced 20% test set to check the model adequency. Then, the model was fine-tuned on the full data. The results on the test set are the following: | | Precision | Recall | F1 | |----------|-----------|--------|-------| | alllang | 0.8713 | 0.8710 | 0.8710| | en | 0.9650 | 0.9650 | 0.9650| | ru | 0.9791 | 0.9790 | 0.9790| | uk | 0.9267 | 0.9250 | 0.9251| | de | 0.8791 | 0.8760 | 0.8758| | es | 0.8700 | 0.8700 | 0.8700| | ar | 0.7787 | 0.7780 | 0.7780| | am | 0.7781 | 0.7780 | 0.7780| | hi | 0.9360 | 0.9360 | 0.9360| | zh | 0.7318 | 0.7320 | 0.7315| Citation If you would like to acknowledge our work, please, cite the following manuscripts:
bert-multilingual-toxicity-classifier
Glot500 Toxicity Classifier
twitter-xlmr-toxicity-classifier
xlmr-base-toxicity-classifier
Llama-pairwise-content-evaluator
This model is a fine-tuned version of meta-llama/Llama-3.1-8B-Instruct. It has been trained using TRL. - TRL: 0.16.0 - Transformers: 4.50.1 - Pytorch: 2.5.1 - Datasets: 3.4.1 - Tokenizers: 0.21.1
Llama-pairwise-toxicity-evaluator
This model is a fine-tuned version of meta-llama/Llama-3.1-8B-Instruct. It has been trained using TRL with textdetox/detoxificationpairwisestyleevaluation dataset - TRL: 0.16.0 - Transformers: 4.50.1 - Pytorch: 2.5.1 - Datasets: 3.4.1 - Tokenizers: 0.21.1