Upload LlamaForCausalLM

#8
by Defetya - opened
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "/kaggle/input/llama2-7b-hf/Llama2-7b-hf",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
@@ -16,7 +16,7 @@
16
  "num_hidden_layers": 32,
17
  "num_key_value_heads": 32,
18
  "pretraining_tp": 1,
19
- "rms_norm_eps": 1e-05,
20
  "rope_scaling": null,
21
  "rope_theta": 10000.0,
22
  "tie_word_embeddings": false,
 
1
  {
2
+ "_name_or_path": "meta-llama/Llama-2-7b-chat-hf",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
16
  "num_hidden_layers": 32,
17
  "num_key_value_heads": 32,
18
  "pretraining_tp": 1,
19
+ "rms_norm_eps": 1e-06,
20
  "rope_scaling": null,
21
  "rope_theta": 10000.0,
22
  "tie_word_embeddings": false,
generation_config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 1,
3
+ "do_sample": true,
4
+ "eos_token_id": 2,
5
+ "max_length": 4096,
6
+ "pad_token_id": 0,
7
+ "temperature": 0.6,
8
+ "top_p": 0.9,
9
+ "transformers_version": "4.34.0"
10
+ }
pytorch_model-00001-of-00002.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7b4baeb3be139292973e3bc293293e8c3a5b4e73c83ec5d0b4a6b7411f3d1e20
3
  size 9976620609
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:784aee31853f3fe3869d431b0ba34a8bc72aee56f47a5a22453869edf985c962
3
  size 9976620609
pytorch_model-00002-of-00002.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6f89f40d3c269d9abe35a3a2fb05d6c03d5681dd1579aa2dcb29448630fd9abe
3
  size 3500311262
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88e284119f3e7164cf3ab68ff21edf5025851e0b51cd4a4609833db787b4063f
3
  size 3500311262
pytorch_model.bin.index.json CHANGED
@@ -295,4 +295,4 @@
295
  "model.layers.9.self_attn.v_proj.weight": "pytorch_model-00001-of-00002.bin",
296
  "model.norm.weight": "pytorch_model-00002-of-00002.bin"
297
  }
298
- }
 
295
  "model.layers.9.self_attn.v_proj.weight": "pytorch_model-00001-of-00002.bin",
296
  "model.norm.weight": "pytorch_model-00002-of-00002.bin"
297
  }
298
+ }