cicdatopea commited on
Commit
9361596
·
verified ·
1 Parent(s): a51c120

Upload config.json with huggingface_hub

Browse files
Files changed (1) hide show
  1. config.json +6 -5
config.json CHANGED
@@ -19,11 +19,11 @@
19
  "quantization_config": {
20
  "amp": true,
21
  "autoround_version": "0.4.5",
22
- "backend": "auto_round:gptq:exllamav2",
23
  "batch_size": 8,
24
  "bits": 4,
 
25
  "data_type": "int",
26
- "dataset": "NeelNanda/pile-10k",
27
  "enable_minmax_tuning": true,
28
  "enable_norm_bias_tuning": false,
29
  "enable_quanted_input": true,
@@ -34,11 +34,12 @@
34
  "lr": 0.005,
35
  "minmax_lr": 0.005,
36
  "nsamples": 128,
37
- "quant_method": "intel/auto-round",
38
  "scale_dtype": "torch.float16",
39
  "seqlen": 2048,
40
  "sym": true,
41
- "to_quant_block_names": null
 
42
  },
43
  "rms_norm_eps": 1e-05,
44
  "rope_scaling": null,
@@ -50,4 +51,4 @@
50
  "use_cache": true,
51
  "use_sliding_window": false,
52
  "vocab_size": 152064
53
- }
 
19
  "quantization_config": {
20
  "amp": true,
21
  "autoround_version": "0.4.5",
 
22
  "batch_size": 8,
23
  "bits": 4,
24
+ "damp_percent": 0.01,
25
  "data_type": "int",
26
+ "desc_act": false,
27
  "enable_minmax_tuning": true,
28
  "enable_norm_bias_tuning": false,
29
  "enable_quanted_input": true,
 
34
  "lr": 0.005,
35
  "minmax_lr": 0.005,
36
  "nsamples": 128,
37
+ "quant_method": "gptq",
38
  "scale_dtype": "torch.float16",
39
  "seqlen": 2048,
40
  "sym": true,
41
+ "to_quant_block_names": null,
42
+ "true_sequential": false
43
  },
44
  "rms_norm_eps": 1e-05,
45
  "rope_scaling": null,
 
51
  "use_cache": true,
52
  "use_sliding_window": false,
53
  "vocab_size": 152064
54
+ }