indukurs commited on
Commit
f2bcfaa
·
1 Parent(s): 3ff18dc

End of training

Browse files
README.md CHANGED
@@ -1,32 +1,13 @@
1
  ---
2
  license: apache-2.0
3
- base_model: indukurs/base-model
4
  tags:
5
  - generated_from_trainer
6
  datasets:
7
  - imdb
8
- metrics:
9
- - accuracy
10
- - f1
11
  model-index:
12
  - name: pruned_model
13
- results:
14
- - task:
15
- name: Text Classification
16
- type: text-classification
17
- dataset:
18
- name: imdb
19
- type: imdb
20
- config: plain_text
21
- split: test
22
- args: plain_text
23
- metrics:
24
- - name: Accuracy
25
- type: accuracy
26
- value: 0.9
27
- - name: F1
28
- type: f1
29
- value: 0.9013157894736842
30
  ---
31
 
32
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -34,11 +15,7 @@ should probably proofread and complete it, then remove this comment. -->
34
 
35
  # pruned_model
36
 
37
- This model is a fine-tuned version of [indukurs/base-model](https://huggingface.co/indukurs/base-model) on the imdb dataset.
38
- It achieves the following results on the evaluation set:
39
- - Loss: 0.4540
40
- - Accuracy: 0.9
41
- - F1: 0.9013
42
 
43
  ## Model description
44
 
 
1
  ---
2
  license: apache-2.0
3
+ base_model: bert-base-uncased
4
  tags:
5
  - generated_from_trainer
6
  datasets:
7
  - imdb
 
 
 
8
  model-index:
9
  - name: pruned_model
10
+ results: []
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
  ---
12
 
13
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
15
 
16
  # pruned_model
17
 
18
+ This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on the imdb dataset.
 
 
 
 
19
 
20
  ## Model description
21
 
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "indukurs/base-model",
3
  "architectures": [
4
  "BertForSequenceClassification"
5
  ],
 
1
  {
2
+ "_name_or_path": "bert-base-uncased",
3
  "architectures": [
4
  "BertForSequenceClassification"
5
  ],
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d18e107858afb180d4611a7fbe9df61a7b68d2aed02161cca83c4ed04e62009b
3
- size 437958648
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:471ab3359ce68584cb824c2e27d4b13ec930e0feff6fd9ce165bfc7639bfdc53
3
+ size 437964888
runs/Nov20_19-29-17_a5268629fa9c/events.out.tfevents.1700509154.a5268629fa9c.2072.3 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:52584a613500af770bee1bd4e14ead7b07fbaf5c243c85eb752201df83b003c6
3
- size 457
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc09142c1cd95a688537fbc6326d99dc689e1276cc9d5cda537338bc7654780b
3
+ size 826
runs/Nov20_19-58-38_a5268629fa9c/events.out.tfevents.1700510320.a5268629fa9c.2072.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61272fa145670f9a0b4ccf28ddb5223c0873198799484fca209a15bbedf0d252
3
+ size 4620
special_tokens_map.json CHANGED
@@ -1,37 +1,7 @@
1
  {
2
- "cls_token": {
3
- "content": "[CLS]",
4
- "lstrip": false,
5
- "normalized": false,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "mask_token": {
10
- "content": "[MASK]",
11
- "lstrip": false,
12
- "normalized": false,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
- "pad_token": {
17
- "content": "[PAD]",
18
- "lstrip": false,
19
- "normalized": false,
20
- "rstrip": false,
21
- "single_word": false
22
- },
23
- "sep_token": {
24
- "content": "[SEP]",
25
- "lstrip": false,
26
- "normalized": false,
27
- "rstrip": false,
28
- "single_word": false
29
- },
30
- "unk_token": {
31
- "content": "[UNK]",
32
- "lstrip": false,
33
- "normalized": false,
34
- "rstrip": false,
35
- "single_word": false
36
- }
37
  }
 
1
  {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
  }
tokenizer_config.json CHANGED
@@ -45,15 +45,11 @@
45
  "cls_token": "[CLS]",
46
  "do_lower_case": true,
47
  "mask_token": "[MASK]",
48
- "max_length": 512,
49
  "model_max_length": 512,
50
  "pad_token": "[PAD]",
51
  "sep_token": "[SEP]",
52
- "stride": 0,
53
  "strip_accents": null,
54
  "tokenize_chinese_chars": true,
55
  "tokenizer_class": "BertTokenizer",
56
- "truncation_side": "right",
57
- "truncation_strategy": "longest_first",
58
  "unk_token": "[UNK]"
59
  }
 
45
  "cls_token": "[CLS]",
46
  "do_lower_case": true,
47
  "mask_token": "[MASK]",
 
48
  "model_max_length": 512,
49
  "pad_token": "[PAD]",
50
  "sep_token": "[SEP]",
 
51
  "strip_accents": null,
52
  "tokenize_chinese_chars": true,
53
  "tokenizer_class": "BertTokenizer",
 
 
54
  "unk_token": "[UNK]"
55
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7f6a21956c3a34a4108c38a9813b36bb0cc8f313432d5263e87d6e28bf5920a7
3
  size 4664
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ca1a83474802d0ba8af01e822df3d5b96f42a6f2282f2412ad37fd47e72fcf3
3
  size 4664