Praful932 commited on
Commit
ff361eb
1 Parent(s): e30a082

Add quantization config

Browse files
Files changed (1) hide show
  1. config.json +9 -1
config.json CHANGED
@@ -23,5 +23,13 @@
23
  "torch_dtype": "float16",
24
  "transformers_version": "4.34.0",
25
  "use_cache": true,
26
- "vocab_size": 32002
 
 
 
 
 
 
 
 
27
  }
 
23
  "torch_dtype": "float16",
24
  "transformers_version": "4.34.0",
25
  "use_cache": true,
26
+ "vocab_size": 32002,
27
+ "quantization_config" : {
28
+ "quant_method": "awq",
29
+ "zero_point": true,
30
+ "group_size": 128,
31
+ "bits": 4,
32
+ "version": "gemm",
33
+ "backend" : "autoawq",
34
+ }
35
  }