sharpenb commited on
Commit
daf100d
·
verified ·
1 Parent(s): 37d6949

Upload folder using huggingface_hub (#3)

Browse files

- 27f9771cf12f313df211f29a5ec5d20bc6364369e1c079fa5926ae9aaa8f987d (2c359ee8746653e6f70bdfb669116fc6df2b5eb1)
- 325169e691bdeccf3dfd2b80266b20bf9c667cd01124727460045259497e00ce (07c8fb2c40a751e79d77098fe267128c003dd7ec)
- 0074c6e4581a76896097919d678b4ac5288839dec4a32e19cf041c8cac2c2f8c (eeaef4a12b0a27a8e4ddda6de72c4d5e19433ec5)

Files changed (2) hide show
  1. config.json +1 -1
  2. smash_config.json +1 -1
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "/covalent/.cache/models/tmp9byiaf56tnw5qgoy",
3
  "architectures": [
4
  "Qwen2ForCausalLM"
5
  ],
 
1
  {
2
+ "_name_or_path": "/covalent/.cache/models/tmpv9fj5d9e8fnd5qlq",
3
  "architectures": [
4
  "Qwen2ForCausalLM"
5
  ],
smash_config.json CHANGED
@@ -28,7 +28,7 @@
28
  "quant_llm-int8_weight_bits": 8,
29
  "max_batch_size": 1,
30
  "device": "cuda",
31
- "cache_dir": "/covalent/.cache/models/tmp9byiaf56",
32
  "task": "",
33
  "save_load_fn": "bitsandbytes",
34
  "save_load_fn_args": {}
 
28
  "quant_llm-int8_weight_bits": 8,
29
  "max_batch_size": 1,
30
  "device": "cuda",
31
+ "cache_dir": "/covalent/.cache/models/tmpv9fj5d9e",
32
  "task": "",
33
  "save_load_fn": "bitsandbytes",
34
  "save_load_fn_args": {}