ntgiaky commited on
Commit
2d8f5c0
·
1 Parent(s): 062066e

Initial upload: PhoBERT NER for Vietnamese Smart Home slot filling

Browse files
added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "<mask>": 64000
3
+ }
bpe.codes ADDED
The diff for this file is too large to render. See raw diff
 
label_mappings.json ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "label2id": {
3
+ "O": 0,
4
+ "B-changing value": 1,
5
+ "I-changing value": 2,
6
+ "B-device": 3,
7
+ "I-device": 4,
8
+ "B-duration": 5,
9
+ "I-duration": 6,
10
+ "B-living_space": 7,
11
+ "I-living_space": 8,
12
+ "B-target number": 9,
13
+ "I-target number": 10,
14
+ "B-time at": 11,
15
+ "I-time at": 12
16
+ },
17
+ "id2label": {
18
+ "0": "O",
19
+ "1": "B-changing value",
20
+ "2": "I-changing value",
21
+ "3": "B-device",
22
+ "4": "I-device",
23
+ "5": "B-duration",
24
+ "6": "I-duration",
25
+ "7": "B-living_space",
26
+ "8": "I-living_space",
27
+ "9": "B-target number",
28
+ "10": "I-target number",
29
+ "11": "B-time at",
30
+ "12": "I-time at"
31
+ },
32
+ "slot_labels": [
33
+ "O",
34
+ "B-changing value",
35
+ "I-changing value",
36
+ "B-device",
37
+ "I-device",
38
+ "B-duration",
39
+ "I-duration",
40
+ "B-living_space",
41
+ "I-living_space",
42
+ "B-target number",
43
+ "I-target number",
44
+ "B-time at",
45
+ "I-time at"
46
+ ]
47
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7434a65d3615865a003237f6feecfb92234c96ed2818e38dd4251e871e27ae5d
3
+ size 537741802
slot_labels.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ "O",
3
+ "B-changing value",
4
+ "I-changing value",
5
+ "B-device",
6
+ "I-device",
7
+ "B-duration",
8
+ "I-duration",
9
+ "B-living_space",
10
+ "I-living_space",
11
+ "B-target number",
12
+ "I-target number",
13
+ "B-time at",
14
+ "I-time at"
15
+ ]
special_tokens_map.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "cls_token": "<s>",
4
+ "eos_token": "</s>",
5
+ "mask_token": "<mask>",
6
+ "pad_token": "<pad>",
7
+ "sep_token": "</s>",
8
+ "unk_token": "<unk>"
9
+ }
tokenizer_config.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "clean_up_tokenization_spaces": true,
4
+ "cls_token": "<s>",
5
+ "eos_token": "</s>",
6
+ "mask_token": "<mask>",
7
+ "model_max_length": 256,
8
+ "pad_token": "<pad>",
9
+ "sep_token": "</s>",
10
+ "tokenizer_class": "PhobertTokenizer",
11
+ "unk_token": "<unk>"
12
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff