OPEA
/

Safetensors
olmo2
4-bit precision
intel/auto-round
wenhuach commited on
Commit
39856e8
1 Parent(s): 4b5e415

change to autoround format

Browse files

Signed-off-by: wenhuach <[email protected]>

config.json CHANGED
@@ -19,11 +19,11 @@
19
  "quantization_config": {
20
  "amp": true,
21
  "autoround_version": "0.4.2.dev",
 
22
  "batch_size": 8,
23
  "bits": 4,
24
- "damp_percent": 0.01,
25
  "data_type": "int",
26
- "desc_act": false,
27
  "enable_minmax_tuning": true,
28
  "enable_norm_bias_tuning": false,
29
  "enable_quanted_input": true,
@@ -34,7 +34,7 @@
34
  "lr": 0.001,
35
  "minmax_lr": 0.001,
36
  "nsamples": 512,
37
- "quant_method": "gptq",
38
  "scale_dtype": "torch.float16",
39
  "seqlen": 2048,
40
  "sym": true,
@@ -81,8 +81,7 @@
81
  "model.layers.38",
82
  "model.layers.39"
83
  ]
84
- ],
85
- "true_sequential": false
86
  },
87
  "rms_norm_eps": 1e-06,
88
  "rope_scaling": null,
 
19
  "quantization_config": {
20
  "amp": true,
21
  "autoround_version": "0.4.2.dev",
22
+ "backend": "auto_round:gptq:exllamav2",
23
  "batch_size": 8,
24
  "bits": 4,
 
25
  "data_type": "int",
26
+ "dataset": "NeelNanda/pile-10k",
27
  "enable_minmax_tuning": true,
28
  "enable_norm_bias_tuning": false,
29
  "enable_quanted_input": true,
 
34
  "lr": 0.001,
35
  "minmax_lr": 0.001,
36
  "nsamples": 512,
37
+ "quant_method": "intel/auto-round",
38
  "scale_dtype": "torch.float16",
39
  "seqlen": 2048,
40
  "sym": true,
 
81
  "model.layers.38",
82
  "model.layers.39"
83
  ]
84
+ ]
 
85
  },
86
  "rms_norm_eps": 1e-06,
87
  "rope_scaling": null,
quantize_config.json → quantization_config.json RENAMED
@@ -60,9 +60,8 @@
60
  ]
61
  ],
62
  "enable_norm_bias_tuning": false,
 
63
  "autoround_version": "0.4.2.dev",
64
- "quant_method": "gptq",
65
- "desc_act": false,
66
- "true_sequential": false,
67
- "damp_percent": 0.01
68
  }
 
60
  ]
61
  ],
62
  "enable_norm_bias_tuning": false,
63
+ "dataset": "NeelNanda/pile-10k",
64
  "autoround_version": "0.4.2.dev",
65
+ "quant_method": "intel/auto-round",
66
+ "backend": "auto_round:gptq:exllamav2"
 
 
67
  }