study-hjt commited on
Commit
d90bc23
1 Parent(s): baec174

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +10 -0
config.json CHANGED
@@ -16,6 +16,16 @@
16
  "num_attention_heads": 64,
17
  "num_hidden_layers": 80,
18
  "num_key_value_heads": 8,
 
 
 
 
 
 
 
 
 
 
19
  "rms_norm_eps": 1e-06,
20
  "rope_theta": 1000000.0,
21
  "sliding_window": 32768,
 
16
  "num_attention_heads": 64,
17
  "num_hidden_layers": 80,
18
  "num_key_value_heads": 8,
19
+ "quantization_config": {
20
+ "bits": 4,
21
+ "damp_percent": 0.1,
22
+ "desc_act": false,
23
+ "group_size": 128,
24
+ "modules_in_block_to_quantize": null,
25
+ "quant_method": "gptq",
26
+ "sym": true,
27
+ "true_sequential": true
28
+ },
29
  "rms_norm_eps": 1e-06,
30
  "rope_theta": 1000000.0,
31
  "sliding_window": 32768,