chansung commited on
Commit
ef59610
·
verified ·
1 Parent(s): 1e98497

Model save

Browse files
README.md CHANGED
@@ -20,7 +20,7 @@ should probably proofread and complete it, then remove this comment. -->
20
 
21
  This model is a fine-tuned version of [google/gemma-7b](https://huggingface.co/google/gemma-7b) on the generator dataset.
22
  It achieves the following results on the evaluation set:
23
- - Loss: 2.8222
24
 
25
  ## Model description
26
 
@@ -57,7 +57,7 @@ The following hyperparameters were used during training:
57
 
58
  | Training Loss | Epoch | Step | Validation Loss |
59
  |:-------------:|:------:|:----:|:---------------:|
60
- | 1.3179 | 0.9977 | 221 | 2.8222 |
61
 
62
 
63
  ### Framework versions
 
20
 
21
  This model is a fine-tuned version of [google/gemma-7b](https://huggingface.co/google/gemma-7b) on the generator dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 2.8123
24
 
25
  ## Model description
26
 
 
57
 
58
  | Training Loss | Epoch | Step | Validation Loss |
59
  |:-------------:|:------:|:----:|:---------------:|
60
+ | 1.3142 | 0.9977 | 221 | 2.8123 |
61
 
62
 
63
  ### Framework versions
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2a146c48f6da46e4999c4af192732c884617f6953364f04986e7b0412e46a0f8
3
  size 12859872
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:330572751706f31589c9562c802bbd155a7191c76cd866552840bf809721303a
3
  size 12859872
all_results.json CHANGED
@@ -1,14 +1,9 @@
1
  {
2
  "epoch": 0.9977426636568849,
3
- "eval_loss": 2.8222014904022217,
4
- "eval_runtime": 1.2855,
5
- "eval_samples": 25,
6
- "eval_samples_per_second": 16.336,
7
- "eval_steps_per_second": 0.778,
8
  "total_flos": 6.738683116954583e+17,
9
- "train_loss": 5.562472593730392,
10
- "train_runtime": 2738.0196,
11
  "train_samples": 129221,
12
- "train_samples_per_second": 10.349,
13
- "train_steps_per_second": 0.081
14
  }
 
1
  {
2
  "epoch": 0.9977426636568849,
 
 
 
 
 
3
  "total_flos": 6.738683116954583e+17,
4
+ "train_loss": 5.577704090877895,
5
+ "train_runtime": 575.1744,
6
  "train_samples": 129221,
7
+ "train_samples_per_second": 49.265,
8
+ "train_steps_per_second": 0.384
9
  }
runs/Nov18_05-23-30_main-lora-gemma7b-summarization-0-0/events.out.tfevents.1731926156.main-lora-gemma7b-summarization-0-0.456.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2dfc8fd61b59b95f5d039092be36ed4be112701dadb4dbd11e406d5e3b8dbc36
3
- size 14260
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:208af1ca59aaf7a9d9779f59cb336c4e796753bc05741fa12223503b7ac1fc72
3
+ size 15729
train_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "epoch": 0.9977426636568849,
3
  "total_flos": 6.738683116954583e+17,
4
- "train_loss": 5.562472593730392,
5
- "train_runtime": 2738.0196,
6
  "train_samples": 129221,
7
- "train_samples_per_second": 10.349,
8
- "train_steps_per_second": 0.081
9
  }
 
1
  {
2
  "epoch": 0.9977426636568849,
3
  "total_flos": 6.738683116954583e+17,
4
+ "train_loss": 5.577704090877895,
5
+ "train_runtime": 575.1744,
6
  "train_samples": 129221,
7
+ "train_samples_per_second": 49.265,
8
+ "train_steps_per_second": 0.384
9
  }
trainer_state.json CHANGED
@@ -10,335 +10,335 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.004514672686230248,
13
- "grad_norm": 200.3717803955078,
14
  "learning_rate": 8.695652173913044e-06,
15
- "loss": 40.073,
16
  "step": 1
17
  },
18
  {
19
  "epoch": 0.022573363431151242,
20
- "grad_norm": 131.19760131835938,
21
  "learning_rate": 4.347826086956522e-05,
22
- "loss": 39.1612,
23
  "step": 5
24
  },
25
  {
26
  "epoch": 0.045146726862302484,
27
- "grad_norm": 41.251068115234375,
28
  "learning_rate": 8.695652173913044e-05,
29
- "loss": 30.2659,
30
  "step": 10
31
  },
32
  {
33
  "epoch": 0.06772009029345373,
34
- "grad_norm": 12.021024703979492,
35
  "learning_rate": 0.00013043478260869567,
36
- "loss": 23.7836,
37
  "step": 15
38
  },
39
  {
40
  "epoch": 0.09029345372460497,
41
- "grad_norm": 13.451888084411621,
42
  "learning_rate": 0.00017391304347826088,
43
- "loss": 20.8756,
44
  "step": 20
45
  },
46
  {
47
  "epoch": 0.11286681715575621,
48
- "grad_norm": 4.8614888191223145,
49
  "learning_rate": 0.00019994965423831854,
50
- "loss": 19.0103,
51
  "step": 25
52
  },
53
  {
54
  "epoch": 0.13544018058690746,
55
- "grad_norm": 8.32628059387207,
56
  "learning_rate": 0.00019938384644612543,
57
- "loss": 18.0406,
58
  "step": 30
59
  },
60
  {
61
  "epoch": 0.1580135440180587,
62
- "grad_norm": 17.078899383544922,
63
  "learning_rate": 0.00019819286972627066,
64
- "loss": 16.4692,
65
  "step": 35
66
  },
67
  {
68
  "epoch": 0.18058690744920994,
69
- "grad_norm": 24.7756404876709,
70
  "learning_rate": 0.00019638421585599423,
71
- "loss": 13.1214,
72
  "step": 40
73
  },
74
  {
75
  "epoch": 0.20316027088036118,
76
- "grad_norm": 31.732440948486328,
77
  "learning_rate": 0.00019396926207859084,
78
- "loss": 8.5547,
79
  "step": 45
80
  },
81
  {
82
  "epoch": 0.22573363431151242,
83
- "grad_norm": 9.871573448181152,
84
  "learning_rate": 0.00019096319953545185,
85
- "loss": 4.2086,
86
  "step": 50
87
  },
88
  {
89
  "epoch": 0.24830699774266365,
90
- "grad_norm": 8.936306953430176,
91
  "learning_rate": 0.00018738493770697852,
92
- "loss": 2.9412,
93
  "step": 55
94
  },
95
  {
96
  "epoch": 0.2708803611738149,
97
- "grad_norm": 3.45522141456604,
98
  "learning_rate": 0.00018325698546347715,
99
- "loss": 2.4618,
100
  "step": 60
101
  },
102
  {
103
  "epoch": 0.29345372460496616,
104
- "grad_norm": 1.241241693496704,
105
  "learning_rate": 0.00017860530947427875,
106
- "loss": 2.1257,
107
  "step": 65
108
  },
109
  {
110
  "epoch": 0.3160270880361174,
111
- "grad_norm": 1.5993455648422241,
112
  "learning_rate": 0.00017345917086575332,
113
- "loss": 1.8936,
114
  "step": 70
115
  },
116
  {
117
  "epoch": 0.33860045146726864,
118
- "grad_norm": 1.037279486656189,
119
  "learning_rate": 0.00016785094115571322,
120
- "loss": 1.7682,
121
  "step": 75
122
  },
123
  {
124
  "epoch": 0.3611738148984199,
125
- "grad_norm": 1.5379009246826172,
126
  "learning_rate": 0.00016181589862206052,
127
- "loss": 1.6873,
128
  "step": 80
129
  },
130
  {
131
  "epoch": 0.3837471783295711,
132
- "grad_norm": 1.3465474843978882,
133
  "learning_rate": 0.00015539200638661104,
134
- "loss": 1.6315,
135
  "step": 85
136
  },
137
  {
138
  "epoch": 0.40632054176072235,
139
- "grad_norm": 0.7901816964149475,
140
  "learning_rate": 0.00014861967361004687,
141
- "loss": 1.5786,
142
  "step": 90
143
  },
144
  {
145
  "epoch": 0.4288939051918736,
146
- "grad_norm": 1.0765341520309448,
147
  "learning_rate": 0.00014154150130018866,
148
- "loss": 1.5464,
149
  "step": 95
150
  },
151
  {
152
  "epoch": 0.45146726862302483,
153
- "grad_norm": 1.1750948429107666,
154
  "learning_rate": 0.00013420201433256689,
155
- "loss": 1.5079,
156
  "step": 100
157
  },
158
  {
159
  "epoch": 0.47404063205417607,
160
- "grad_norm": 1.1422131061553955,
161
  "learning_rate": 0.00012664738136900348,
162
- "loss": 1.488,
163
  "step": 105
164
  },
165
  {
166
  "epoch": 0.4966139954853273,
167
- "grad_norm": 1.0460481643676758,
168
  "learning_rate": 0.00011892512443604102,
169
- "loss": 1.4541,
170
  "step": 110
171
  },
172
  {
173
  "epoch": 0.5191873589164786,
174
- "grad_norm": 1.787310242652893,
175
  "learning_rate": 0.00011108381999010111,
176
- "loss": 1.4428,
177
  "step": 115
178
  },
179
  {
180
  "epoch": 0.5417607223476298,
181
- "grad_norm": 0.8201248645782471,
182
  "learning_rate": 0.00010317279334980678,
183
- "loss": 1.4187,
184
  "step": 120
185
  },
186
  {
187
  "epoch": 0.5643340857787811,
188
- "grad_norm": 0.8423493504524231,
189
  "learning_rate": 9.524180841762577e-05,
190
- "loss": 1.4086,
191
  "step": 125
192
  },
193
  {
194
  "epoch": 0.5869074492099323,
195
- "grad_norm": 2.127373218536377,
196
  "learning_rate": 8.734075464262507e-05,
197
- "loss": 1.3947,
198
  "step": 130
199
  },
200
  {
201
  "epoch": 0.6094808126410836,
202
- "grad_norm": 0.8410171270370483,
203
  "learning_rate": 7.951933319348095e-05,
204
- "loss": 1.3942,
205
  "step": 135
206
  },
207
  {
208
  "epoch": 0.6320541760722348,
209
- "grad_norm": 0.7370259761810303,
210
  "learning_rate": 7.182674431585704e-05,
211
- "loss": 1.3738,
212
  "step": 140
213
  },
214
  {
215
  "epoch": 0.654627539503386,
216
- "grad_norm": 0.6766214370727539,
217
  "learning_rate": 6.431137784081282e-05,
218
- "loss": 1.3752,
219
  "step": 145
220
  },
221
  {
222
  "epoch": 0.6772009029345373,
223
- "grad_norm": 0.7337061762809753,
224
  "learning_rate": 5.702050879108284e-05,
225
- "loss": 1.3607,
226
  "step": 150
227
  },
228
  {
229
  "epoch": 0.6997742663656885,
230
- "grad_norm": 0.835713267326355,
231
  "learning_rate": 5.000000000000002e-05,
232
- "loss": 1.3486,
233
  "step": 155
234
  },
235
  {
236
  "epoch": 0.7223476297968398,
237
- "grad_norm": 0.7193433046340942,
238
  "learning_rate": 4.329401361372294e-05,
239
- "loss": 1.3472,
240
  "step": 160
241
  },
242
  {
243
  "epoch": 0.744920993227991,
244
- "grad_norm": 0.8834360837936401,
245
  "learning_rate": 3.694473329154778e-05,
246
- "loss": 1.3397,
247
  "step": 165
248
  },
249
  {
250
  "epoch": 0.7674943566591422,
251
- "grad_norm": 0.9133020639419556,
252
  "learning_rate": 3.099209885178882e-05,
253
- "loss": 1.3359,
254
  "step": 170
255
  },
256
  {
257
  "epoch": 0.7900677200902935,
258
- "grad_norm": 0.6195141673088074,
259
  "learning_rate": 2.5473555032424533e-05,
260
- "loss": 1.3314,
261
  "step": 175
262
  },
263
  {
264
  "epoch": 0.8126410835214447,
265
- "grad_norm": 0.6716246604919434,
266
  "learning_rate": 2.042381594691678e-05,
267
- "loss": 1.3376,
268
  "step": 180
269
  },
270
  {
271
  "epoch": 0.835214446952596,
272
- "grad_norm": 0.8972969055175781,
273
  "learning_rate": 1.587464671688187e-05,
274
- "loss": 1.3334,
275
  "step": 185
276
  },
277
  {
278
  "epoch": 0.8577878103837472,
279
- "grad_norm": 0.6790732145309448,
280
  "learning_rate": 1.1854663655241805e-05,
281
- "loss": 1.3305,
282
  "step": 190
283
  },
284
  {
285
  "epoch": 0.8803611738148984,
286
- "grad_norm": 0.7163174748420715,
287
  "learning_rate": 8.38915425679304e-06,
288
- "loss": 1.3258,
289
  "step": 195
290
  },
291
  {
292
  "epoch": 0.9029345372460497,
293
- "grad_norm": 0.7202246189117432,
294
  "learning_rate": 5.499918128533155e-06,
295
- "loss": 1.33,
296
  "step": 200
297
  },
298
  {
299
  "epoch": 0.9255079006772009,
300
- "grad_norm": 0.6578797101974487,
301
  "learning_rate": 3.2051298603643753e-06,
302
- "loss": 1.3235,
303
  "step": 205
304
  },
305
  {
306
  "epoch": 0.9480812641083521,
307
- "grad_norm": 0.6833933591842651,
308
  "learning_rate": 1.5192246987791981e-06,
309
- "loss": 1.327,
310
  "step": 210
311
  },
312
  {
313
  "epoch": 0.9706546275395034,
314
- "grad_norm": 1.3088898658752441,
315
  "learning_rate": 4.5280774269154115e-07,
316
- "loss": 1.3266,
317
  "step": 215
318
  },
319
  {
320
  "epoch": 0.9932279909706546,
321
- "grad_norm": 0.6681888699531555,
322
  "learning_rate": 1.2587232612493172e-08,
323
- "loss": 1.3179,
324
  "step": 220
325
  },
326
  {
327
  "epoch": 0.9977426636568849,
328
- "eval_loss": 2.8222014904022217,
329
- "eval_runtime": 1.269,
330
- "eval_samples_per_second": 16.549,
331
- "eval_steps_per_second": 0.788,
332
  "step": 221
333
  },
334
  {
335
  "epoch": 0.9977426636568849,
336
  "step": 221,
337
  "total_flos": 6.738683116954583e+17,
338
- "train_loss": 5.562472593730392,
339
- "train_runtime": 2738.0196,
340
- "train_samples_per_second": 10.349,
341
- "train_steps_per_second": 0.081
342
  }
343
  ],
344
  "logging_steps": 5,
 
10
  "log_history": [
11
  {
12
  "epoch": 0.004514672686230248,
13
+ "grad_norm": 199.82510375976562,
14
  "learning_rate": 8.695652173913044e-06,
15
+ "loss": 39.0472,
16
  "step": 1
17
  },
18
  {
19
  "epoch": 0.022573363431151242,
20
+ "grad_norm": 178.4315185546875,
21
  "learning_rate": 4.347826086956522e-05,
22
+ "loss": 39.3436,
23
  "step": 5
24
  },
25
  {
26
  "epoch": 0.045146726862302484,
27
+ "grad_norm": 40.531532287597656,
28
  "learning_rate": 8.695652173913044e-05,
29
+ "loss": 30.4834,
30
  "step": 10
31
  },
32
  {
33
  "epoch": 0.06772009029345373,
34
+ "grad_norm": 12.844618797302246,
35
  "learning_rate": 0.00013043478260869567,
36
+ "loss": 23.9376,
37
  "step": 15
38
  },
39
  {
40
  "epoch": 0.09029345372460497,
41
+ "grad_norm": 15.092730522155762,
42
  "learning_rate": 0.00017391304347826088,
43
+ "loss": 20.9382,
44
  "step": 20
45
  },
46
  {
47
  "epoch": 0.11286681715575621,
48
+ "grad_norm": 4.834229469299316,
49
  "learning_rate": 0.00019994965423831854,
50
+ "loss": 19.0458,
51
  "step": 25
52
  },
53
  {
54
  "epoch": 0.13544018058690746,
55
+ "grad_norm": 8.221763610839844,
56
  "learning_rate": 0.00019938384644612543,
57
+ "loss": 18.0883,
58
  "step": 30
59
  },
60
  {
61
  "epoch": 0.1580135440180587,
62
+ "grad_norm": 16.106189727783203,
63
  "learning_rate": 0.00019819286972627066,
64
+ "loss": 16.5366,
65
  "step": 35
66
  },
67
  {
68
  "epoch": 0.18058690744920994,
69
+ "grad_norm": 24.50678825378418,
70
  "learning_rate": 0.00019638421585599423,
71
+ "loss": 13.2306,
72
  "step": 40
73
  },
74
  {
75
  "epoch": 0.20316027088036118,
76
+ "grad_norm": 32.74522399902344,
77
  "learning_rate": 0.00019396926207859084,
78
+ "loss": 8.7225,
79
  "step": 45
80
  },
81
  {
82
  "epoch": 0.22573363431151242,
83
+ "grad_norm": 9.876773834228516,
84
  "learning_rate": 0.00019096319953545185,
85
+ "loss": 4.3489,
86
  "step": 50
87
  },
88
  {
89
  "epoch": 0.24830699774266365,
90
+ "grad_norm": 8.977824211120605,
91
  "learning_rate": 0.00018738493770697852,
92
+ "loss": 2.9443,
93
  "step": 55
94
  },
95
  {
96
  "epoch": 0.2708803611738149,
97
+ "grad_norm": 3.1251025199890137,
98
  "learning_rate": 0.00018325698546347715,
99
+ "loss": 2.4464,
100
  "step": 60
101
  },
102
  {
103
  "epoch": 0.29345372460496616,
104
+ "grad_norm": 1.4109467267990112,
105
  "learning_rate": 0.00017860530947427875,
106
+ "loss": 2.0921,
107
  "step": 65
108
  },
109
  {
110
  "epoch": 0.3160270880361174,
111
+ "grad_norm": 1.3768062591552734,
112
  "learning_rate": 0.00017345917086575332,
113
+ "loss": 1.8687,
114
  "step": 70
115
  },
116
  {
117
  "epoch": 0.33860045146726864,
118
+ "grad_norm": 1.025557279586792,
119
  "learning_rate": 0.00016785094115571322,
120
+ "loss": 1.7492,
121
  "step": 75
122
  },
123
  {
124
  "epoch": 0.3611738148984199,
125
+ "grad_norm": 1.717323660850525,
126
  "learning_rate": 0.00016181589862206052,
127
+ "loss": 1.6714,
128
  "step": 80
129
  },
130
  {
131
  "epoch": 0.3837471783295711,
132
+ "grad_norm": 1.101912021636963,
133
  "learning_rate": 0.00015539200638661104,
134
+ "loss": 1.6189,
135
  "step": 85
136
  },
137
  {
138
  "epoch": 0.40632054176072235,
139
+ "grad_norm": 0.8903200030326843,
140
  "learning_rate": 0.00014861967361004687,
141
+ "loss": 1.5678,
142
  "step": 90
143
  },
144
  {
145
  "epoch": 0.4288939051918736,
146
+ "grad_norm": 1.1036828756332397,
147
  "learning_rate": 0.00014154150130018866,
148
+ "loss": 1.5342,
149
  "step": 95
150
  },
151
  {
152
  "epoch": 0.45146726862302483,
153
+ "grad_norm": 1.4335256814956665,
154
  "learning_rate": 0.00013420201433256689,
155
+ "loss": 1.496,
156
  "step": 100
157
  },
158
  {
159
  "epoch": 0.47404063205417607,
160
+ "grad_norm": 1.3628050088882446,
161
  "learning_rate": 0.00012664738136900348,
162
+ "loss": 1.4813,
163
  "step": 105
164
  },
165
  {
166
  "epoch": 0.4966139954853273,
167
+ "grad_norm": 1.7230000495910645,
168
  "learning_rate": 0.00011892512443604102,
169
+ "loss": 1.4475,
170
  "step": 110
171
  },
172
  {
173
  "epoch": 0.5191873589164786,
174
+ "grad_norm": 2.127577066421509,
175
  "learning_rate": 0.00011108381999010111,
176
+ "loss": 1.4365,
177
  "step": 115
178
  },
179
  {
180
  "epoch": 0.5417607223476298,
181
+ "grad_norm": 1.139145851135254,
182
  "learning_rate": 0.00010317279334980678,
183
+ "loss": 1.413,
184
  "step": 120
185
  },
186
  {
187
  "epoch": 0.5643340857787811,
188
+ "grad_norm": 1.1161638498306274,
189
  "learning_rate": 9.524180841762577e-05,
190
+ "loss": 1.4018,
191
  "step": 125
192
  },
193
  {
194
  "epoch": 0.5869074492099323,
195
+ "grad_norm": 1.6872923374176025,
196
  "learning_rate": 8.734075464262507e-05,
197
+ "loss": 1.3883,
198
  "step": 130
199
  },
200
  {
201
  "epoch": 0.6094808126410836,
202
+ "grad_norm": 1.0828768014907837,
203
  "learning_rate": 7.951933319348095e-05,
204
+ "loss": 1.3903,
205
  "step": 135
206
  },
207
  {
208
  "epoch": 0.6320541760722348,
209
+ "grad_norm": 1.0485174655914307,
210
  "learning_rate": 7.182674431585704e-05,
211
+ "loss": 1.3705,
212
  "step": 140
213
  },
214
  {
215
  "epoch": 0.654627539503386,
216
+ "grad_norm": 0.8391625285148621,
217
  "learning_rate": 6.431137784081282e-05,
218
+ "loss": 1.3695,
219
  "step": 145
220
  },
221
  {
222
  "epoch": 0.6772009029345373,
223
+ "grad_norm": 0.7003734707832336,
224
  "learning_rate": 5.702050879108284e-05,
225
+ "loss": 1.3572,
226
  "step": 150
227
  },
228
  {
229
  "epoch": 0.6997742663656885,
230
+ "grad_norm": 0.9220963716506958,
231
  "learning_rate": 5.000000000000002e-05,
232
+ "loss": 1.3447,
233
  "step": 155
234
  },
235
  {
236
  "epoch": 0.7223476297968398,
237
+ "grad_norm": 0.8875526189804077,
238
  "learning_rate": 4.329401361372294e-05,
239
+ "loss": 1.3435,
240
  "step": 160
241
  },
242
  {
243
  "epoch": 0.744920993227991,
244
+ "grad_norm": 1.013006329536438,
245
  "learning_rate": 3.694473329154778e-05,
246
+ "loss": 1.3352,
247
  "step": 165
248
  },
249
  {
250
  "epoch": 0.7674943566591422,
251
+ "grad_norm": 0.7449764609336853,
252
  "learning_rate": 3.099209885178882e-05,
253
+ "loss": 1.3322,
254
  "step": 170
255
  },
256
  {
257
  "epoch": 0.7900677200902935,
258
+ "grad_norm": 0.7195131778717041,
259
  "learning_rate": 2.5473555032424533e-05,
260
+ "loss": 1.3275,
261
  "step": 175
262
  },
263
  {
264
  "epoch": 0.8126410835214447,
265
+ "grad_norm": 0.6962352395057678,
266
  "learning_rate": 2.042381594691678e-05,
267
+ "loss": 1.3348,
268
  "step": 180
269
  },
270
  {
271
  "epoch": 0.835214446952596,
272
+ "grad_norm": 0.8499621152877808,
273
  "learning_rate": 1.587464671688187e-05,
274
+ "loss": 1.3293,
275
  "step": 185
276
  },
277
  {
278
  "epoch": 0.8577878103837472,
279
+ "grad_norm": 0.6392541527748108,
280
  "learning_rate": 1.1854663655241805e-05,
281
+ "loss": 1.325,
282
  "step": 190
283
  },
284
  {
285
  "epoch": 0.8803611738148984,
286
+ "grad_norm": 0.7239265441894531,
287
  "learning_rate": 8.38915425679304e-06,
288
+ "loss": 1.3204,
289
  "step": 195
290
  },
291
  {
292
  "epoch": 0.9029345372460497,
293
+ "grad_norm": 0.7242350578308105,
294
  "learning_rate": 5.499918128533155e-06,
295
+ "loss": 1.324,
296
  "step": 200
297
  },
298
  {
299
  "epoch": 0.9255079006772009,
300
+ "grad_norm": 0.6468668580055237,
301
  "learning_rate": 3.2051298603643753e-06,
302
+ "loss": 1.3198,
303
  "step": 205
304
  },
305
  {
306
  "epoch": 0.9480812641083521,
307
+ "grad_norm": 0.6733763813972473,
308
  "learning_rate": 1.5192246987791981e-06,
309
+ "loss": 1.3226,
310
  "step": 210
311
  },
312
  {
313
  "epoch": 0.9706546275395034,
314
+ "grad_norm": 1.3773088455200195,
315
  "learning_rate": 4.5280774269154115e-07,
316
+ "loss": 1.3211,
317
  "step": 215
318
  },
319
  {
320
  "epoch": 0.9932279909706546,
321
+ "grad_norm": 0.689209520816803,
322
  "learning_rate": 1.2587232612493172e-08,
323
+ "loss": 1.3142,
324
  "step": 220
325
  },
326
  {
327
  "epoch": 0.9977426636568849,
328
+ "eval_loss": 2.81227445602417,
329
+ "eval_runtime": 0.459,
330
+ "eval_samples_per_second": 45.75,
331
+ "eval_steps_per_second": 2.179,
332
  "step": 221
333
  },
334
  {
335
  "epoch": 0.9977426636568849,
336
  "step": 221,
337
  "total_flos": 6.738683116954583e+17,
338
+ "train_loss": 5.577704090877895,
339
+ "train_runtime": 575.1744,
340
+ "train_samples_per_second": 49.265,
341
+ "train_steps_per_second": 0.384
342
  }
343
  ],
344
  "logging_steps": 5,