thanhduc1180 commited on
Commit
18bc77d
1 Parent(s): 36451a0

Training in progress, step 200

Browse files
adapter_config.json CHANGED
@@ -20,8 +20,8 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "q_proj",
24
- "v_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "v_proj",
24
+ "q_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8881586bb9566aefef3c2c3f378b975987a068dffd6f504f9e7c5ae306ed7a3e
3
  size 109098378
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49cd355fcd6188f49ca0966c9b1048ba7997c059e9cb32e3e2967fd247ed8396
3
  size 109098378
special_tokens_map.json CHANGED
@@ -6,6 +6,6 @@
6
  ],
7
  "bos_token": "<s>",
8
  "eos_token": "</s>",
9
- "pad_token": "<unk>",
10
  "unk_token": "<unk>"
11
  }
 
6
  ],
7
  "bos_token": "<s>",
8
  "eos_token": "</s>",
9
+ "pad_token": "</s>",
10
  "unk_token": "<unk>"
11
  }
tokenizer_config.json CHANGED
@@ -69,7 +69,7 @@
69
  "legacy": true,
70
  "max_length": 1024,
71
  "model_max_length": 1000000000000000019884624838656,
72
- "pad_token": "<unk>",
73
  "sp_model_kwargs": {},
74
  "spaces_between_special_tokens": false,
75
  "stride": 0,
 
69
  "legacy": true,
70
  "max_length": 1024,
71
  "model_max_length": 1000000000000000019884624838656,
72
+ "pad_token": "</s>",
73
  "sp_model_kwargs": {},
74
  "spaces_between_special_tokens": false,
75
  "stride": 0,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:983f85f301f430ad9d99fa710c6449c96127b336102a08c3c9b845988bb01c8f
3
  size 4536
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef1646240e4141014de3dd31f3f57bf92dac0b6db0b3dea462760ffd6d555da9
3
  size 4536