YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

Email-spam-detection

This model detects whether an email message is spam or not spam (ham) using a fine-tuned transformer-based classifier.

Model Details

Model Description

This is a binary text classification model trained to distinguish spam emails from legitimate (ham) emails. The model is based on a pretrained transformer architecture (e.g., BERT, RoBERTa) and fine-tuned on a labeled email dataset containing both spam and non-spam messages.

  • Model type: Transformer-based binary classifier
  • Language(s) (NLP): English
  • License: MIT License
  • Finetuned from model : bert-base-uncased (example)

Model Sources [optional]

Loading the Model

from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch

tokenizer = AutoTokenizer.from_pretrained("deepak/email-spam-detection")
model = AutoModelForSequenceClassification.from_pretrained("deepak/email-spam-detection")

emails = [
    "Congratulations! You have won a $1000 gift card. Click here to claim.",
    "Meeting moved to 3 PM today in the conference room.",
]

inputs = tokenizer(emails, padding=True, truncation=True, return_tensors="pt")
outputs = model(**inputs)
logits = outputs.logits
predictions = torch.argmax(logits, dim=-1)
print(predictions)  # 1 = spam, 0 = ham

Training Details

The model was trained on a labeled dataset of emails combining public spam corpora such as the Enron Spam dataset and other sources, balanced between spam and ham emails. Data preprocessing included cleaning email text, removing metadata, and tokenization.

Training Procedure

Emails were normalized by removing special characters and tokenized using the pretrained tokenizer.

Training Hyperparameters

Training regime: fine-tuning with fp16 mixed precision on NVIDIA GPUs

  • Batch size: 32

  • Learning rate: 2e-5

  • Epochs: 4

  • Speeds, Sizes, Times [optional]

  • Checkpoint size: ~400MB

  • Training time: ~3 hours on 1 GPU

Evaluation

Testing Data, Factors & Metrics Testing Data Evaluation was performed on a held-out test split from the same dataset, containing unseen emails.

Factors

  • No explicit subpopulation disaggregation.

Metrics

  • Accuracy

  • Precision

  • Recall

  • F1-score

Results

Metric : Score Accuracy : 0.95 Precision : 0.93 Recall : 0.92 F1-score : 0.925

Model Examination

Attention analysis indicates the model focuses on key spam indicators like suspicious URLs, urgent calls to action, and financial keywords.

Downloads last month
2
Safetensors
Model size
109M params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support