shuyuej commited on
Commit
49ad6b4
1 Parent(s): 5400942

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -18
README.md CHANGED
@@ -10,28 +10,17 @@ Link: [https://huggingface.co/meta-llama/Meta-Llama-3.1-70B-Instruct](https://hu
10
  ## Quantization Configurations
11
  ```
12
  "quantization_config": {
13
- "batch_size": 1,
14
  "bits": 4,
15
- "block_name_to_quantize": null,
16
- "cache_block_outputs": true,
17
- "damp_percent": 0.1,
18
- "dataset": null,
19
- "desc_act": false,
20
- "exllama_config": {
21
- "version": 1
22
- },
23
  "group_size": 128,
24
- "max_input_length": null,
25
- "model_seqlen": null,
26
- "module_name_preceding_first_block": null,
27
- "modules_in_block_to_quantize": null,
28
- "pad_token_id": null,
29
  "quant_method": "gptq",
 
30
  "sym": true,
31
- "tokenizer": null,
32
- "true_sequential": true,
33
- "use_cuda_fp16": false,
34
- "use_exllama": true
35
  },
36
  ```
37
 
 
10
  ## Quantization Configurations
11
  ```
12
  "quantization_config": {
 
13
  "bits": 4,
14
+ "checkpoint_format": "gptq",
15
+ "damp_percent": 0.01,
16
+ "desc_act": true,
 
 
 
 
 
17
  "group_size": 128,
18
+ "model_file_base_name": null,
19
+ "model_name_or_path": null,
 
 
 
20
  "quant_method": "gptq",
21
+ "static_groups": false,
22
  "sym": true,
23
+ "true_sequential": true
 
 
 
24
  },
25
  ```
26