ViktorDo commited on
Commit
95804ef
·
1 Parent(s): 2d21809

Upload 2 files

Browse files
Files changed (2) hide show
  1. config.json +1 -22
  2. vocab.txt +0 -0
config.json CHANGED
@@ -1,22 +1 @@
1
- {
2
- "activation": "gelu",
3
- "architectures": [
4
- "DistilBertForMaskedLM"
5
- ],
6
- "attention_dropout": 0.1,
7
- "dim": 768,
8
- "dropout": 0.1,
9
- "hidden_dim": 3072,
10
- "initializer_range": 0.02,
11
- "max_position_embeddings": 124,
12
- "model_type": "distilbert",
13
- "n_heads": 12,
14
- "n_layers": 6,
15
- "pad_token_id": 0,
16
- "qa_dropout": 0.1,
17
- "seq_classif_dropout": 0.2,
18
- "sinusoidal_pos_embds": false,
19
- "torch_dtype": "float32",
20
- "transformers_version": "4.26.1",
21
- "vocab_size": 30522
22
- }
 
1
+ {"do_lower_case": true, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "model_max_length": 124, "max_len": 124}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
vocab.txt ADDED
The diff for this file is too large to render. See raw diff