Cheng98 commited on
Commit
5b8af69
·
verified ·
1 Parent(s): d0e2dcb

Add files using upload-large-folder tool

Browse files
meta-llama-Llama-2-7b-hf-mxint8/token-1024-pos-1/formatted_tensors.json CHANGED
@@ -36,6 +36,16 @@
36
  },
37
  "hex": "saved_tensors/meta-llama-Llama-2-7b-hf-mxint8/token-1024-pos-1/meta-llama_Llama-2-7b-hf.model.layers.0.input_layernorm.input.npy"
38
  },
 
 
 
 
 
 
 
 
 
 
39
  "meta-llama_Llama-2-7b-hf.model.layers.0.input_layernorm.output": {
40
  "tensor_meta": {
41
  "is_emulated": false,
@@ -420,6 +430,16 @@
420
  },
421
  "hex": "saved_tensors/meta-llama-Llama-2-7b-hf-mxint8/token-1024-pos-1/meta-llama_Llama-2-7b-hf.model.layers.0.post_attention_layernorm.input.npy"
422
  },
 
 
 
 
 
 
 
 
 
 
423
  "meta-llama_Llama-2-7b-hf.model.layers.0.post_attention_layernorm.output": {
424
  "tensor_meta": {
425
  "is_emulated": false,
@@ -657,6 +677,16 @@
657
  },
658
  "hex": "saved_tensors/meta-llama-Llama-2-7b-hf-mxint8/token-1024-pos-1/meta-llama_Llama-2-7b-hf.model.norm.input.npy"
659
  },
 
 
 
 
 
 
 
 
 
 
660
  "meta-llama_Llama-2-7b-hf.model.norm.output": {
661
  "tensor_meta": {
662
  "is_emulated": false,
 
36
  },
37
  "hex": "saved_tensors/meta-llama-Llama-2-7b-hf-mxint8/token-1024-pos-1/meta-llama_Llama-2-7b-hf.model.layers.0.input_layernorm.input.npy"
38
  },
39
+ "meta-llama_Llama-2-7b-hf.model.layers.0.input_layernorm.weight": {
40
+ "tensor_meta": {
41
+ "is_emulated": false,
42
+ "dtype": "torch.bfloat16",
43
+ "shape": [
44
+ 4096
45
+ ]
46
+ },
47
+ "hex": "saved_tensors/meta-llama-Llama-2-7b-hf-mxint8/token-1024-pos-1/meta-llama_Llama-2-7b-hf.model.layers.0.input_layernorm.weight.npy"
48
+ },
49
  "meta-llama_Llama-2-7b-hf.model.layers.0.input_layernorm.output": {
50
  "tensor_meta": {
51
  "is_emulated": false,
 
430
  },
431
  "hex": "saved_tensors/meta-llama-Llama-2-7b-hf-mxint8/token-1024-pos-1/meta-llama_Llama-2-7b-hf.model.layers.0.post_attention_layernorm.input.npy"
432
  },
433
+ "meta-llama_Llama-2-7b-hf.model.layers.0.post_attention_layernorm.weight": {
434
+ "tensor_meta": {
435
+ "is_emulated": false,
436
+ "dtype": "torch.bfloat16",
437
+ "shape": [
438
+ 4096
439
+ ]
440
+ },
441
+ "hex": "saved_tensors/meta-llama-Llama-2-7b-hf-mxint8/token-1024-pos-1/meta-llama_Llama-2-7b-hf.model.layers.0.post_attention_layernorm.weight.npy"
442
+ },
443
  "meta-llama_Llama-2-7b-hf.model.layers.0.post_attention_layernorm.output": {
444
  "tensor_meta": {
445
  "is_emulated": false,
 
677
  },
678
  "hex": "saved_tensors/meta-llama-Llama-2-7b-hf-mxint8/token-1024-pos-1/meta-llama_Llama-2-7b-hf.model.norm.input.npy"
679
  },
680
+ "meta-llama_Llama-2-7b-hf.model.norm.weight": {
681
+ "tensor_meta": {
682
+ "is_emulated": false,
683
+ "dtype": "torch.bfloat16",
684
+ "shape": [
685
+ 4096
686
+ ]
687
+ },
688
+ "hex": "saved_tensors/meta-llama-Llama-2-7b-hf-mxint8/token-1024-pos-1/meta-llama_Llama-2-7b-hf.model.norm.weight.npy"
689
+ },
690
  "meta-llama_Llama-2-7b-hf.model.norm.output": {
691
  "tensor_meta": {
692
  "is_emulated": false,
meta-llama-Llama-2-7b-hf-mxint8/token-1024-pos-1/meta-llama_Llama-2-7b-hf.model.layers.0.input_layernorm.weight.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae53ed863403da64c65fa4d8c80221e02702475685a39b56a316a2bcc71c084b
3
+ size 65664
meta-llama-Llama-2-7b-hf-mxint8/token-1024-pos-1/meta-llama_Llama-2-7b-hf.model.layers.0.post_attention_layernorm.weight.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4dd1ec470dc05006e5310d90279ec3a0e99bd45ed4cce589ccd26218654f6fe9
3
+ size 65664
meta-llama-Llama-2-7b-hf-mxint8/token-1024-pos-1/meta-llama_Llama-2-7b-hf.model.norm.weight.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d0e272ddf00a89945a062ecc7446e9c49919eed1bef04d7756e00ca21a8ca8c
3
+ size 65664