Model Card for MisileLab/noMoreSpamYT
A transformer-based model for detecting bot-generated spam comments on YouTube, with a focus on Korean content promoting adult content and gambling websites.
Model Details
Model Description
noMoreSpamYT is a fine-tuned KcELECTRA model designed to identify and filter bot comments on YouTube videos. It specifically targets automated comments that promote adult content or gambling websites using repetitive patterns and specific keywords in Korean. The model uses a combination of CLS token and mean pooling strategies with custom classification layers to achieve high accuracy in distinguishing between human and bot-generated content.
- Developed by: MisileLab
- Model type: Fine-tuned KcELECTRA for sequence classification
- Language(s) (NLP): Korean (ko)
- License: MIT
- Finetuned from model: KcELECTRA
Model Sources [optional]
- Repository: https://github.com/misilelab/noMoreSpam
- Train code & result: https://static.marimo.app/static/nomorespam-zvfn
Uses
Direct Use
This model is suitable for:
- Detecting spam bot comments in Korean YouTube content
- Filtering promotional comments for adult content and gambling websites
- Content moderation systems for Korean social media platforms
- Research on automated spam detection in Korean text
Downstream Use [optional]
The model can be integrated into:
- YouTube comment moderation systems
- Content filtering pipelines for Korean platforms
- Research frameworks studying bot behavior and spam patterns
- Social media monitoring tools
Out-of-Scope Use
This model should not be used for:
- General text classification tasks unrelated to spam detection
- Detection of sophisticated bots beyond the patterns it was trained on
- Applications requiring high precision in non-Korean languages
- Making decisions about content without human review
- Censorship of legitimate speech or opinions
Bias, Risks, and Limitations
- Pattern dependency: The model relies on specific keywords and patterns that may become outdated
- Language specificity: Optimized for Korean language and may not work well for other languages
- Bot type limitation: Focuses specifically on adult/gambling promotion bots, not all spam types
- Temporal relevance: Bot patterns evolve over time, potentially reducing long-term effectiveness
- False positives: Legitimate comments containing flagged keywords may be misclassified
- Domain specificity: Trained on YouTube comments which may not transfer well to other platforms
Recommendations
Users (both direct and downstream) should:
- Regularly update the model as spam techniques evolve
- Use in combination with other detection methods for robust spam filtering
- Consider both precision and recall when evaluating performance
- Ensure human review of flagged content before taking action
- Monitor for evolving bot patterns and retrain the model periodically
- Be aware of potential biases in the training data
How to Get Started with the Model
Use the code below to get started with the model.
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch
# Load model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("MisileLab/noMoreSpamYT")
model = AutoModelForSequenceClassification.from_pretrained("MisileLab/noMoreSpamYT")
# Prepare input
comment = "여기 방문하세요 19금 즐거움이 가득합니다" # Example spam comment
inputs = tokenizer(comment, return_tensors="pt", padding=True, truncation=True, max_length=512)
# Make prediction
with torch.no_grad():
outputs = model(**inputs)
predictions = torch.nn.functional.softmax(outputs.logits, dim=-1)
is_bot = predictions[0][1].item() > 0.5
probability = predictions[0][1].item()
print(f"Is bot comment: {is_bot}, Probability: {probability:.4f}")
Training Details
Training Data
The model was trained on the youtube-bot-comments-v2 dataset, which contains:
- 50% human comments, 50% bot comments (balanced dataset)
- Comments collected from top South Korean YouTube videos
- Manual and regex-based classification
- Focus on identifying repetitive promotional patterns for adult and gambling websites
Training Procedure
Preprocessing [optional]
- Comments were tokenized using the KcELECTRA tokenizer
- Texts were truncated to a maximum length of 512 tokens
- Data was split into training (80%) and validation (20%) sets
- Special tokens were added for classification
Training Hyperparameters
- Training regime: fp16 mixed precision
- Optimizer: AdamW
- Learning rate: 2e-5
- Batch size: 16
- Epochs: 3
- Loss function: Focal Loss (to handle class imbalance)
- Early stopping: Based on validation F1 score
- Weight decay: 0.01
- Warmup steps: 500
Evaluation
Testing Data, Factors & Metrics
Testing Data
The model was evaluated on a held-out test set (20%) from the youtube-bot-comments-v2 dataset.
Metrics
- Precision: Measures the proportion of predicted bot comments that are actually bots
- Recall: Measures the proportion of actual bot comments that were correctly identified
- F1 Score: Harmonic mean of precision and recall
- Accuracy: Overall proportion of correct predictions
Results
Summary
- Precision: 1
- Recall: 1
- F1 Score: 1
- Accuracy: 1
The model performs well on Korean YouTube comments, particularly for detecting common spam patterns promoting adult content and gambling websites.
Model Examination [optional]
Attention visualization shows the model focuses heavily on specific Korean keywords and patterns associated with spam content, such as "19금" (adult content indicator), gambling-related terms, and URL patterns.
Technical Specifications [optional]
Model Architecture and Objective
The model architecture includes:
- Base KcELECTRA transformer with frozen initial layers
- Custom classification head with:
- Dropout layers (rate=0.1) for regularization
- Combined CLS token and mean pooling strategy
- Two fully connected layers with GELU activation
- Binary classification output with sigmoid activation
Citation [optional]
BibTeX:
@misc{misile2025nomorespam,
title={noMoreSpam: Korean YouTube Bot Comment Detection Model},
author={MisileLab},
year={2025},
howpublished={\url{https://huggingface.co/MisileLab/noMoreSpamYT}}
}
APA:
MisileLab. (2025). noMoreSpam: Korean YouTube Bot Comment Detection Model. https://huggingface.co/MisileLab/noMoreSpamYT
Glossary [optional]
- KcELECTRA: Korean-centric ELECTRA model, a transformer-based model pre-trained on Korean text
- Bot comment: Automated comment typically promoting adult content or gambling websites
- Focal Loss: A loss function that addresses class imbalance by focusing on hard examples
- CLS token: Special classification token used in transformer models for sequence classification
More Information [optional]
For more information about the project and its development, visit:
Model Card Contact
For questions or issues regarding this model, please contact Misile ([email protected]).
Model tree for MisileLab/noMoreSpamYT
Base model
beomi/KcELECTRA-base