sharpenb commited on
Commit
5a17cb6
·
verified ·
1 Parent(s): 8d2a8ee

Upload folder using huggingface_hub (#3)

Browse files

- fdb53d2e90c3da8df0fe9f0b23e033952548242f6ee5c69ef041c6a196307d56 (2bef3962b40d9c7a8b16b64d0e816b2a942090e3)
- 121ec7baa904570547b3fa34cb5efdd7e98abb5283a6270d2219154c25eaa10b (c00beae2b54fe84374fb6320dc8ebb176b9ec543)
- 8cae25045aa27337dafb37b78cbe02fadbc78167291c52bdad59cda956993549 (eb38baa59b499a20e8a8362c4df738c1dc3db246)

Files changed (2) hide show
  1. config.json +1 -1
  2. smash_config.json +37 -0
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "/covalent/.cache/models/tmp905d1b9pt94gnbuk",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
1
  {
2
+ "_name_or_path": "/covalent/.cache/models/tmp6cox54sy70vwnjyh",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
smash_config.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "comp_cgenerate_active": false,
3
+ "comp_ctranslate_active": false,
4
+ "comp_cwhisper_active": false,
5
+ "comp_diffusers2_active": false,
6
+ "comp_ifw_active": false,
7
+ "comp_onediff_active": false,
8
+ "comp_step_caching_active": false,
9
+ "comp_torch_compile_active": false,
10
+ "comp_ws2t_active": false,
11
+ "comp_x-fast_active": false,
12
+ "prune_torch-structured_active": false,
13
+ "quant_aqlm_active": false,
14
+ "quant_awq_active": false,
15
+ "quant_gptq_active": false,
16
+ "quant_half_active": false,
17
+ "quant_hqq_active": false,
18
+ "quant_llm-int8_active": true,
19
+ "quant_quanto_active": false,
20
+ "quant_torch_dynamic_active": false,
21
+ "quant_torch_static_active": false,
22
+ "quant_llm-int8_compute_dtype": "bfloat16",
23
+ "quant_llm-int8_double_quant": false,
24
+ "quant_llm-int8_enable_fp32_cpu_offload": false,
25
+ "quant_llm-int8_has_fp16_weight": false,
26
+ "quant_llm-int8_quant_type": "fp4",
27
+ "quant_llm-int8_threshold": 6.0,
28
+ "quant_llm-int8_weight_bits": 8,
29
+ "max_batch_size": 1,
30
+ "device": "cuda",
31
+ "cache_dir": "/covalent/.cache/models/tmp6cox54sy",
32
+ "task": "",
33
+ "save_load_fn": "bitsandbytes",
34
+ "save_load_fn_args": {
35
+ "weight_quantization_bits": "param.dtype"
36
+ }
37
+ }