leejunhyeok commited on
Commit
37ed692
·
verified ·
1 Parent(s): 83d1f84

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +0 -5
config.json CHANGED
@@ -18,7 +18,6 @@
18
  "max_position_embeddings": 16384,
19
  "max_window_layers": 28,
20
  "model_type": "Motif",
21
- "norm_topk_prob": null,
22
  "num_attention_heads": 16,
23
  "num_hidden_layers": 32,
24
  "num_key_value_heads": 16,
@@ -27,14 +26,10 @@
27
  "rope_theta": 500000.0,
28
  "sliding_window": null,
29
  "tie_word_embeddings": true,
30
- "topk_group": null,
31
- "topk_method": null,
32
  "torch_dtype": "float32",
33
  "transformers_version": "4.46.3",
34
  "use_bias": false,
35
  "use_cache": true,
36
- "use_qk_norm": false,
37
  "use_sliding_window": false,
38
  "vocab_size": 219520,
39
- "wesar_weights": false
40
  }
 
18
  "max_position_embeddings": 16384,
19
  "max_window_layers": 28,
20
  "model_type": "Motif",
 
21
  "num_attention_heads": 16,
22
  "num_hidden_layers": 32,
23
  "num_key_value_heads": 16,
 
26
  "rope_theta": 500000.0,
27
  "sliding_window": null,
28
  "tie_word_embeddings": true,
 
 
29
  "torch_dtype": "float32",
30
  "transformers_version": "4.46.3",
31
  "use_bias": false,
32
  "use_cache": true,
 
33
  "use_sliding_window": false,
34
  "vocab_size": 219520,
 
35
  }