File size: 1,660 Bytes
2d6a713
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
---
license: mit
language: en
library_name: onnxruntime
pipeline_tag: text-classification
tags:
- roberta
- spam
- text-classification
- onnx
- distilled
- quantized
base_model: mshenoda/roberta-spam
---

# ONNX Distilled Spam Classifier

This repository contains a distilled and quantized version of a RoBERTa-based spam classification model, optimized for high-performance CPU inference in the ONNX format.

This model was created by distilling `mshenoda/roberta-spam` for the purpose of efficient on-device and cross-platform deployment.

## Model Description

* **Model Type:** A distilled RoBERTa-base model.
* **Task:** Spam classification (binary classification).
* **Format:** ONNX, with dynamic quantization.
* **Key Features:** Lightweight, fast, and ideal for CPU-based inference.

## Intended Uses & Limitations

This model is designed for client-side applications where performance and low resource usage are critical. It's perfect for:

* Desktop applications (Windows, Linux, macOS)
* Mobile applications (with an appropriate ONNX runtime)
* Edge devices

As a distilled model, there may be a minor trade-off in accuracy compared to the larger `roberta-base` teacher model, in exchange for a significant boost in speed and a smaller memory footprint.

## How to Get Started

You can use this model directly with the `onnxruntime` and `transformers` libraries.

### 1. Installation

First, make sure you have the necessary libraries installed. For GPU usage, install `onnxruntime-gpu`; for CPU-only, `onnxruntime` is sufficient.

```bash
# For CPU
pip install onnxruntime transformers

# OR for NVIDIA GPU
pip install onnxruntime-gpu transformers