This model is part of the work done in .
The full code can be found at https://github.com/wetey/cluster-errors

Model Details

Model Description

  • Model type: Distil-BERT
  • Language(s) (NLP): English
  • Finetuned from model: distilbert-base-uncased

How to Get Started with the Model

Use the code below to get started with the model.

# Use a pipeline as a high-level helper
from transformers import pipeline

pipe = pipeline("text-classification", model="wetey/distilbert-base-uncased-measuring-hate-speech")
# Load model directly
from transformers import AutoTokenizer, AutoModelForSequenceClassification

tokenizer = AutoTokenizer.from_pretrained("wetey/distilbert-base-uncased-measuring-hate-speech")
model = AutoModelForSequenceClassification.from_pretrained("wetey/distilbert-base-uncased-measuring-hate-speech")

Fine-tuning Details

Fine-tuning Data

The model was fine-tuned on the ucberkeley-dlab/measuring-hate-speech dataset.
We converted the continuous hatespeech scores to categorical labels using the ranges suggested by the authors. The ranges are listed on the HuggingFace Dataset card.
Examples with hatespeech scores that are lower than -1 are considered supportive, between -1 and 0.5 are neutral, and scores greater than 0.5 are hatespeech.
We remove duplicate examples along with those that received fewer than three total annotations, and we drop the neutral class.
After these steps, we were left with 12,289 examples with 7497 examples labeled as supportive and 4792 labeled as hatespeech. We use 85% of the dataset for fine-tuning and 15% for testing.

Fine-tuning Procedure

The exact fine-tuning procedure followed can be found here

Fine-tuning Hyperparameters

evaluation_strategy = 'epoch'
logging_steps = 1,
num_train_epochs = 5,
learning_rate = 1e-5,
eval_accumulation_steps = 2

Evaluation

Testing Data

Test set used can be found here

Results

accuracy: 89.3%
precision: 89.4%
recall: 89.3%
f1-score: 89.3%

Results per class

Label Precision Recall F1-score
supportive 92% 91% 91%
hatespeech 86% 87% 86%

Citation

Downloads last month
1
Safetensors
Model size
67M params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for wetey/distilbert-base-uncased-measuring-hate-speech

Finetuned
(8073)
this model

Dataset used to train wetey/distilbert-base-uncased-measuring-hate-speech