AlekseyKorshuk commited on
Commit
45accb9
1 Parent(s): 58a9e3b

huggingartists

Browse files
README.md CHANGED
@@ -45,15 +45,15 @@ from datasets import load_dataset
45
  dataset = load_dataset("huggingartists/xxxtentacion")
46
  ```
47
 
48
- [Explore the data](https://wandb.ai/huggingartists/huggingartists/runs/wyqwtyke/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline.
49
 
50
  ## Training procedure
51
 
52
  The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on XXXTENTACION's lyrics.
53
 
54
- Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/huggingartists/huggingartists/runs/2x3eplag) for full transparency and reproducibility.
55
 
56
- At the end of training, [the final model](https://wandb.ai/huggingartists/huggingartists/runs/2x3eplag/artifacts) is logged and versioned.
57
 
58
  ## How to use
59
 
 
45
  dataset = load_dataset("huggingartists/xxxtentacion")
46
  ```
47
 
48
+ [Explore the data](https://wandb.ai/huggingartists/huggingartists/runs/1fgk6duz/artifacts), which is tracked with [W&B artifacts](https://docs.wandb.com/artifacts) at every step of the pipeline.
49
 
50
  ## Training procedure
51
 
52
  The model is based on a pre-trained [GPT-2](https://huggingface.co/gpt2) which is fine-tuned on XXXTENTACION's lyrics.
53
 
54
+ Hyperparameters and metrics are recorded in the [W&B training run](https://wandb.ai/huggingartists/huggingartists/runs/7g46vrtf) for full transparency and reproducibility.
55
 
56
+ At the end of training, [the final model](https://wandb.ai/huggingartists/huggingartists/runs/7g46vrtf/artifacts) is logged and versioned.
57
 
58
  ## How to use
59
 
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "gpt2",
3
  "activation_function": "gelu_new",
4
  "architectures": [
5
  "GPT2LMHeadModel"
@@ -35,7 +35,7 @@
35
  }
36
  },
37
  "torch_dtype": "float32",
38
- "transformers_version": "4.9.2",
39
  "use_cache": true,
40
  "vocab_size": 50257
41
  }
 
1
  {
2
+ "_name_or_path": "huggingartists/xxxtentacion",
3
  "activation_function": "gelu_new",
4
  "architectures": [
5
  "GPT2LMHeadModel"
 
35
  }
36
  },
37
  "torch_dtype": "float32",
38
+ "transformers_version": "4.10.0",
39
  "use_cache": true,
40
  "vocab_size": 50257
41
  }
evaluation.txt CHANGED
@@ -1 +1 @@
1
- {"eval_loss": 3.047952175140381, "eval_runtime": 6.9475, "eval_samples_per_second": 21.015, "eval_steps_per_second": 2.735, "epoch": 1.0}
 
1
+ {"eval_loss": 2.998302936553955, "eval_runtime": 7.549, "eval_samples_per_second": 21.062, "eval_steps_per_second": 2.649, "epoch": 2.0}
flax_model.msgpack CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e31086314908b688e94dea0204e120631a3c14a2a6cca0183f14233865948b8b
3
  size 497764120
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:473c6ce96484afb22b5f44b9754253982be214dde5ccfc6e4636c65d85527beb
3
  size 497764120
optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:57fded92ce9908006c8c0c695c50b6321ba18e3a1898a7ae1ece83f466b0420a
3
- size 995603825
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42035e763018344e9c12b57960c5622bcb42c42ca8545f22d5745457bc43b7f0
3
+ size 995604017
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3e27ba50977e8aa3f88b1483e1ff32e00f8e54b9b104e94626aaaf4109400399
3
  size 510403817
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b9bc902bf8c5fd91ac6140769cbcb1ed04fda15388d328a739a52c62d0e2b7c4
3
  size 510403817
rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d08c08d76d72721a447ad4765dc955cf5c332927d2829e29207096ce695ec488
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3691a8950b6ce0a7fbcdee171960f1672959fb5e260f352ba982df89dd517c41
3
  size 14503
scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a5bc99df78269e2d0dbaca15016af30dea6ca5b37f97a8df9318aec4d0d9c188
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0816ea5e3b1d810ba13a63e3db58b55d3131231050f4d5136f720f83ed0afbf
3
  size 623
tokenizer_config.json CHANGED
@@ -1 +1 @@
1
- {"unk_token": "<|endoftext|>", "bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "add_prefix_space": false, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "gpt2", "tokenizer_class": "GPT2Tokenizer"}
 
1
+ {"unk_token": "<|endoftext|>", "bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "add_prefix_space": false, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "huggingartists/xxxtentacion", "tokenizer_class": "GPT2Tokenizer"}
trainer_state.json CHANGED
@@ -1,156 +1,302 @@
1
  {
2
- "best_metric": 3.047952175140381,
3
- "best_model_checkpoint": "output/xxxtentacion/checkpoint-114",
4
- "epoch": 1.0,
5
- "global_step": 114,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
9
  "log_history": [
10
  {
11
  "epoch": 0.04,
12
- "learning_rate": 0.00013654981500613273,
13
- "loss": 3.7924,
14
  "step": 5
15
  },
16
  {
17
  "epoch": 0.09,
18
- "learning_rate": 0.00013461158482121457,
19
- "loss": 3.5933,
20
  "step": 10
21
  },
22
  {
23
  "epoch": 0.13,
24
- "learning_rate": 0.00013142205020853694,
25
- "loss": 3.4387,
26
  "step": 15
27
  },
28
  {
29
  "epoch": 0.18,
30
- "learning_rate": 0.00012704167144630924,
31
- "loss": 3.3335,
32
  "step": 20
33
  },
34
  {
35
  "epoch": 0.22,
36
- "learning_rate": 0.00012155348225285961,
37
- "loss": 3.3589,
38
  "step": 25
39
  },
40
  {
41
- "epoch": 0.26,
42
- "learning_rate": 0.00011506151581352585,
43
- "loss": 3.3227,
44
  "step": 30
45
  },
46
  {
47
  "epoch": 0.31,
48
- "learning_rate": 0.00010768883274520426,
49
- "loss": 3.6526,
50
  "step": 35
51
  },
52
  {
53
  "epoch": 0.35,
54
- "learning_rate": 9.957518838012702e-05,
55
- "loss": 3.2754,
56
  "step": 40
57
  },
58
  {
59
- "epoch": 0.39,
60
- "learning_rate": 9.08743835874413e-05,
61
- "loss": 3.2572,
62
  "step": 45
63
  },
64
  {
65
  "epoch": 0.44,
66
- "learning_rate": 8.175134934996847e-05,
67
- "loss": 3.1639,
68
  "step": 50
69
  },
70
  {
71
- "epoch": 0.48,
72
- "learning_rate": 7.237902036041238e-05,
73
- "loss": 3.343,
74
  "step": 55
75
  },
76
  {
77
  "epoch": 0.53,
78
- "learning_rate": 6.293505690059801e-05,
79
- "loss": 3.4637,
80
  "step": 60
81
  },
82
  {
83
- "epoch": 0.57,
84
- "learning_rate": 5.359847714324262e-05,
85
- "loss": 3.1047,
86
  "step": 65
87
  },
88
  {
89
- "epoch": 0.61,
90
- "learning_rate": 4.4546263713777056e-05,
91
- "loss": 3.0869,
92
  "step": 70
93
  },
94
  {
95
  "epoch": 0.66,
96
- "learning_rate": 3.5950008837656755e-05,
97
- "loss": 3.3149,
98
  "step": 75
99
  },
100
  {
101
- "epoch": 0.7,
102
- "learning_rate": 2.7972661667182378e-05,
103
- "loss": 3.107,
104
  "step": 80
105
  },
106
  {
107
  "epoch": 0.75,
108
- "learning_rate": 2.0765439444951916e-05,
109
- "loss": 2.9627,
110
  "step": 85
111
  },
112
  {
113
- "epoch": 0.79,
114
- "learning_rate": 1.4464961055407408e-05,
115
- "loss": 3.0099,
116
  "step": 90
117
  },
118
  {
119
- "epoch": 0.83,
120
- "learning_rate": 9.190657300387505e-06,
121
- "loss": 3.5708,
122
  "step": 95
123
  },
124
  {
125
  "epoch": 0.88,
126
- "learning_rate": 5.042506989064599e-06,
127
- "loss": 3.0849,
128
  "step": 100
129
  },
130
  {
131
- "epoch": 0.92,
132
- "learning_rate": 2.0991417565619363e-06,
133
- "loss": 3.1334,
134
  "step": 105
135
  },
136
  {
137
- "epoch": 0.96,
138
- "learning_rate": 4.1635553598866533e-07,
139
- "loss": 3.2816,
140
  "step": 110
141
  },
142
  {
143
  "epoch": 1.0,
144
- "eval_loss": 3.047952175140381,
145
- "eval_runtime": 6.9391,
146
- "eval_samples_per_second": 21.04,
147
- "eval_steps_per_second": 2.738,
148
- "step": 114
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
149
  }
150
  ],
151
- "max_steps": 114,
152
- "num_train_epochs": 1,
153
- "total_flos": 118887874560000.0,
154
  "trial_name": null,
155
  "trial_params": null
156
  }
 
1
  {
2
+ "best_metric": 2.998302936553955,
3
+ "best_model_checkpoint": "output/xxxtentacion/checkpoint-226",
4
+ "epoch": 2.0,
5
+ "global_step": 226,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
9
  "log_history": [
10
  {
11
  "epoch": 0.04,
12
+ "learning_rate": 0.0001365382750064078,
13
+ "loss": 2.8729,
14
  "step": 5
15
  },
16
  {
17
  "epoch": 0.09,
18
+ "learning_rate": 0.00013456586620543126,
19
+ "loss": 2.8615,
20
  "step": 10
21
  },
22
  {
23
  "epoch": 0.13,
24
+ "learning_rate": 0.0001313208258478339,
25
+ "loss": 2.8047,
26
  "step": 15
27
  },
28
  {
29
  "epoch": 0.18,
30
+ "learning_rate": 0.00012686575814089965,
31
+ "loss": 2.9533,
32
  "step": 20
33
  },
34
  {
35
  "epoch": 0.22,
36
+ "learning_rate": 0.00012128661147092064,
37
+ "loss": 2.8864,
38
  "step": 25
39
  },
40
  {
41
+ "epoch": 0.27,
42
+ "learning_rate": 0.00011469102026339097,
43
+ "loss": 2.5404,
44
  "step": 30
45
  },
46
  {
47
  "epoch": 0.31,
48
+ "learning_rate": 0.00010720622847019214,
49
+ "loss": 3.0059,
50
  "step": 35
51
  },
52
  {
53
  "epoch": 0.35,
54
+ "learning_rate": 9.897663474442597e-05,
55
+ "loss": 2.827,
56
  "step": 40
57
  },
58
  {
59
+ "epoch": 0.4,
60
+ "learning_rate": 9.01610066620532e-05,
61
+ "loss": 2.804,
62
  "step": 45
63
  },
64
  {
65
  "epoch": 0.44,
66
+ "learning_rate": 8.092941773433672e-05,
67
+ "loss": 2.8347,
68
  "step": 50
69
  },
70
  {
71
+ "epoch": 0.49,
72
+ "learning_rate": 7.14599663030852e-05,
73
+ "loss": 2.7375,
74
  "step": 55
75
  },
76
  {
77
  "epoch": 0.53,
78
+ "learning_rate": 6.193533961866806e-05,
79
+ "loss": 2.5601,
80
  "step": 60
81
  },
82
  {
83
+ "epoch": 0.58,
84
+ "learning_rate": 5.253928938754926e-05,
85
+ "loss": 2.8287,
86
  "step": 65
87
  },
88
  {
89
+ "epoch": 0.62,
90
+ "learning_rate": 4.345308678404285e-05,
91
+ "loss": 2.7017,
92
  "step": 70
93
  },
94
  {
95
  "epoch": 0.66,
96
+ "learning_rate": 3.4852025317174834e-05,
97
+ "loss": 2.7319,
98
  "step": 75
99
  },
100
  {
101
+ "epoch": 0.71,
102
+ "learning_rate": 2.6902039020299325e-05,
103
+ "loss": 2.7863,
104
  "step": 80
105
  },
106
  {
107
  "epoch": 0.75,
108
+ "learning_rate": 1.97565012062839e-05,
109
+ "loss": 2.7227,
110
  "step": 85
111
  },
112
  {
113
+ "epoch": 0.8,
114
+ "learning_rate": 1.3553265547553997e-05,
115
+ "loss": 2.4648,
116
  "step": 90
117
  },
118
  {
119
+ "epoch": 0.84,
120
+ "learning_rate": 8.412006565291007e-06,
121
+ "loss": 2.4789,
122
  "step": 95
123
  },
124
  {
125
  "epoch": 0.88,
126
+ "learning_rate": 4.431910835795193e-06,
127
+ "loss": 2.6082,
128
  "step": 100
129
  },
130
  {
131
+ "epoch": 0.93,
132
+ "learning_rate": 1.6897634558914594e-06,
133
+ "loss": 2.8574,
134
  "step": 105
135
  },
136
  {
137
+ "epoch": 0.97,
138
+ "learning_rate": 2.384666838129088e-07,
139
+ "loss": 2.5917,
140
  "step": 110
141
  },
142
  {
143
  "epoch": 1.0,
144
+ "eval_loss": 3.0134646892547607,
145
+ "eval_runtime": 7.5402,
146
+ "eval_samples_per_second": 21.087,
147
+ "eval_steps_per_second": 2.652,
148
+ "step": 113
149
+ },
150
+ {
151
+ "epoch": 1.02,
152
+ "learning_rate": 1.0601933434684325e-07,
153
+ "loss": 2.4012,
154
+ "step": 115
155
+ },
156
+ {
157
+ "epoch": 1.06,
158
+ "learning_rate": 1.2949766180319454e-06,
159
+ "loss": 2.5116,
160
+ "step": 120
161
+ },
162
+ {
163
+ "epoch": 1.11,
164
+ "learning_rate": 3.7824008453368197e-06,
165
+ "loss": 2.682,
166
+ "step": 125
167
+ },
168
+ {
169
+ "epoch": 1.15,
170
+ "learning_rate": 7.5203039468480955e-06,
171
+ "loss": 2.6505,
172
+ "step": 130
173
+ },
174
+ {
175
+ "epoch": 1.19,
176
+ "learning_rate": 1.2436573272248243e-05,
177
+ "loss": 2.5776,
178
+ "step": 135
179
+ },
180
+ {
181
+ "epoch": 1.24,
182
+ "learning_rate": 1.8436362807314313e-05,
183
+ "loss": 2.631,
184
+ "step": 140
185
+ },
186
+ {
187
+ "epoch": 1.28,
188
+ "learning_rate": 2.540392296919934e-05,
189
+ "loss": 2.4606,
190
+ "step": 145
191
+ },
192
+ {
193
+ "epoch": 1.33,
194
+ "learning_rate": 3.320483367908481e-05,
195
+ "loss": 2.3774,
196
+ "step": 150
197
+ },
198
+ {
199
+ "epoch": 1.37,
200
+ "learning_rate": 4.1688597631154055e-05,
201
+ "loss": 2.5596,
202
+ "step": 155
203
+ },
204
+ {
205
+ "epoch": 1.42,
206
+ "learning_rate": 5.06915437278321e-05,
207
+ "loss": 2.5551,
208
+ "step": 160
209
+ },
210
+ {
211
+ "epoch": 1.46,
212
+ "learning_rate": 6.003998466742405e-05,
213
+ "loss": 2.4846,
214
+ "step": 165
215
+ },
216
+ {
217
+ "epoch": 1.5,
218
+ "learning_rate": 6.955356776710644e-05,
219
+ "loss": 2.5243,
220
+ "step": 170
221
+ },
222
+ {
223
+ "epoch": 1.55,
224
+ "learning_rate": 7.904875437627624e-05,
225
+ "loss": 2.5396,
226
+ "step": 175
227
+ },
228
+ {
229
+ "epoch": 1.59,
230
+ "learning_rate": 8.834236075446375e-05,
231
+ "loss": 2.4629,
232
+ "step": 180
233
+ },
234
+ {
235
+ "epoch": 1.64,
236
+ "learning_rate": 9.725509210222169e-05,
237
+ "loss": 2.5566,
238
+ "step": 185
239
+ },
240
+ {
241
+ "epoch": 1.68,
242
+ "learning_rate": 0.00010561500156549765,
243
+ "loss": 2.5873,
244
+ "step": 190
245
+ },
246
+ {
247
+ "epoch": 1.73,
248
+ "learning_rate": 0.00011326080748143412,
249
+ "loss": 2.5575,
250
+ "step": 195
251
+ },
252
+ {
253
+ "epoch": 1.77,
254
+ "learning_rate": 0.00012004500486837714,
255
+ "loss": 2.3709,
256
+ "step": 200
257
+ },
258
+ {
259
+ "epoch": 1.81,
260
+ "learning_rate": 0.0001258367111323791,
261
+ "loss": 2.6089,
262
+ "step": 205
263
+ },
264
+ {
265
+ "epoch": 1.86,
266
+ "learning_rate": 0.0001305241910900454,
267
+ "loss": 2.5663,
268
+ "step": 210
269
+ },
270
+ {
271
+ "epoch": 1.9,
272
+ "learning_rate": 0.00013401701259429296,
273
+ "loss": 2.6963,
274
+ "step": 215
275
+ },
276
+ {
277
+ "epoch": 1.95,
278
+ "learning_rate": 0.00013624779117609787,
279
+ "loss": 2.6391,
280
+ "step": 220
281
+ },
282
+ {
283
+ "epoch": 1.99,
284
+ "learning_rate": 0.0001371734900441264,
285
+ "loss": 2.7449,
286
+ "step": 225
287
+ },
288
+ {
289
+ "epoch": 2.0,
290
+ "eval_loss": 2.998302936553955,
291
+ "eval_runtime": 7.5393,
292
+ "eval_samples_per_second": 21.089,
293
+ "eval_steps_per_second": 2.653,
294
+ "step": 226
295
  }
296
  ],
297
+ "max_steps": 226,
298
+ "num_train_epochs": 2,
299
+ "total_flos": 234378952704000.0,
300
  "trial_name": null,
301
  "trial_params": null
302
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e74c5392a6e0e5419a3c8e54cc2a6e2abd8a96dda1e8be3831c74bd250130fe2
3
  size 2671
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:82f2723ca1e95981d792cb45af7783a851ea2dee4eed340a95ef82a8fb6a5a29
3
  size 2671