chansung commited on
Commit
0ddf2b6
·
verified ·
1 Parent(s): 532525a

Model save

Browse files
README.md CHANGED
@@ -20,7 +20,7 @@ should probably proofread and complete it, then remove this comment. -->
20
 
21
  This model is a fine-tuned version of [meta-llama/Meta-Llama-3-8B](https://huggingface.co/meta-llama/Meta-Llama-3-8B) on the generator dataset.
22
  It achieves the following results on the evaluation set:
23
- - Loss: 2.3717
24
 
25
  ## Model description
26
 
@@ -57,7 +57,7 @@ The following hyperparameters were used during training:
57
 
58
  | Training Loss | Epoch | Step | Validation Loss |
59
  |:-------------:|:------:|:----:|:---------------:|
60
- | 1.6835 | 0.9955 | 111 | 2.3717 |
61
 
62
 
63
  ### Framework versions
 
20
 
21
  This model is a fine-tuned version of [meta-llama/Meta-Llama-3-8B](https://huggingface.co/meta-llama/Meta-Llama-3-8B) on the generator dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 2.3661
24
 
25
  ## Model description
26
 
 
57
 
58
  | Training Loss | Epoch | Step | Validation Loss |
59
  |:-------------:|:------:|:----:|:---------------:|
60
+ | 1.6751 | 0.9955 | 111 | 2.3661 |
61
 
62
 
63
  ### Framework versions
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3db40600a5aac974df42c1a3419e32e028e994dd039e1229ad5f1f88d9b2cf73
3
  size 4216407520
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce3f5f0cd52569dc2224ccc17fed04275c06a42123e9aa446a2d55388fe90c3c
3
  size 4216407520
all_results.json CHANGED
@@ -1,14 +1,9 @@
1
  {
2
  "epoch": 0.9955156950672646,
3
- "eval_loss": 2.371652841567993,
4
- "eval_runtime": 1.1184,
5
- "eval_samples": 25,
6
- "eval_samples_per_second": 18.776,
7
- "eval_steps_per_second": 0.894,
8
  "total_flos": 6.554327132594504e+17,
9
- "train_loss": 1.8387699009061933,
10
- "train_runtime": 1914.8973,
11
  "train_samples": 129221,
12
- "train_samples_per_second": 14.857,
13
- "train_steps_per_second": 0.058
14
  }
 
1
  {
2
  "epoch": 0.9955156950672646,
 
 
 
 
 
3
  "total_flos": 6.554327132594504e+17,
4
+ "train_loss": 1.8339371971181921,
5
+ "train_runtime": 536.3339,
6
  "train_samples": 129221,
7
+ "train_samples_per_second": 53.043,
8
+ "train_steps_per_second": 0.207
9
  }
runs/Nov18_05-54-20_main-lora-mistral7b-summarization-0-0/events.out.tfevents.1731928130.main-lora-mistral7b-summarization-0-0.456.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:532f840d6a2a6445f2f01a7634a28cddacd42118d87cd8a219c64d415332d8c0
3
- size 9914
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2728a122a208c53c5acc6c77d83dffdf9d7014176ef9b2874c7a67ec4d7afbc
3
+ size 10942
train_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "epoch": 0.9955156950672646,
3
  "total_flos": 6.554327132594504e+17,
4
- "train_loss": 1.8387699009061933,
5
- "train_runtime": 1914.8973,
6
  "train_samples": 129221,
7
- "train_samples_per_second": 14.857,
8
- "train_steps_per_second": 0.058
9
  }
 
1
  {
2
  "epoch": 0.9955156950672646,
3
  "total_flos": 6.554327132594504e+17,
4
+ "train_loss": 1.8339371971181921,
5
+ "train_runtime": 536.3339,
6
  "train_samples": 129221,
7
+ "train_samples_per_second": 53.043,
8
+ "train_steps_per_second": 0.207
9
  }
trainer_state.json CHANGED
@@ -10,181 +10,181 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.008968609865470852,
13
- "grad_norm": 1.1094273328781128,
14
  "learning_rate": 1.6666666666666667e-05,
15
- "loss": 2.4851,
16
  "step": 1
17
  },
18
  {
19
  "epoch": 0.04484304932735426,
20
- "grad_norm": 0.9243674278259277,
21
  "learning_rate": 8.333333333333334e-05,
22
- "loss": 2.4898,
23
  "step": 5
24
  },
25
  {
26
  "epoch": 0.08968609865470852,
27
- "grad_norm": 0.6846497654914856,
28
  "learning_rate": 0.0001666666666666667,
29
- "loss": 2.4067,
30
  "step": 10
31
  },
32
  {
33
  "epoch": 0.13452914798206278,
34
- "grad_norm": 0.8646454811096191,
35
  "learning_rate": 0.00019954719225730847,
36
- "loss": 2.2437,
37
  "step": 15
38
  },
39
  {
40
  "epoch": 0.17937219730941703,
41
- "grad_norm": 0.659982442855835,
42
  "learning_rate": 0.00019679487013963564,
43
- "loss": 2.0573,
44
  "step": 20
45
  },
46
  {
47
  "epoch": 0.2242152466367713,
48
- "grad_norm": 0.579827070236206,
49
  "learning_rate": 0.00019161084574320696,
50
- "loss": 1.9415,
51
  "step": 25
52
  },
53
  {
54
  "epoch": 0.26905829596412556,
55
- "grad_norm": 0.3958710730075836,
56
  "learning_rate": 0.00018412535328311814,
57
- "loss": 1.868,
58
  "step": 30
59
  },
60
  {
61
  "epoch": 0.31390134529147984,
62
- "grad_norm": 0.2995990514755249,
63
  "learning_rate": 0.0001745264449675755,
64
- "loss": 1.814,
65
  "step": 35
66
  },
67
  {
68
  "epoch": 0.35874439461883406,
69
- "grad_norm": 0.2901516556739807,
70
  "learning_rate": 0.00016305526670845226,
71
- "loss": 1.7888,
72
  "step": 40
73
  },
74
  {
75
  "epoch": 0.40358744394618834,
76
- "grad_norm": 0.27101069688796997,
77
  "learning_rate": 0.00015000000000000001,
78
- "loss": 1.759,
79
  "step": 45
80
  },
81
  {
82
  "epoch": 0.4484304932735426,
83
- "grad_norm": 0.252288281917572,
84
  "learning_rate": 0.00013568862215918717,
85
- "loss": 1.7476,
86
  "step": 50
87
  },
88
  {
89
  "epoch": 0.49327354260089684,
90
- "grad_norm": 0.24804328382015228,
91
  "learning_rate": 0.00012048066680651908,
92
- "loss": 1.738,
93
  "step": 55
94
  },
95
  {
96
  "epoch": 0.5381165919282511,
97
- "grad_norm": 0.25213855504989624,
98
  "learning_rate": 0.00010475819158237425,
99
- "loss": 1.7199,
100
  "step": 60
101
  },
102
  {
103
  "epoch": 0.5829596412556054,
104
- "grad_norm": 0.25072190165519714,
105
  "learning_rate": 8.891618000989891e-05,
106
- "loss": 1.7053,
107
  "step": 65
108
  },
109
  {
110
  "epoch": 0.6278026905829597,
111
- "grad_norm": 0.24229365587234497,
112
  "learning_rate": 7.335261863099651e-05,
113
- "loss": 1.708,
114
  "step": 70
115
  },
116
  {
117
  "epoch": 0.672645739910314,
118
- "grad_norm": 0.245011568069458,
119
  "learning_rate": 5.845849869981137e-05,
120
- "loss": 1.6943,
121
  "step": 75
122
  },
123
  {
124
  "epoch": 0.7174887892376681,
125
- "grad_norm": 0.25127002596855164,
126
  "learning_rate": 4.4607993613388976e-05,
127
- "loss": 1.6854,
128
  "step": 80
129
  },
130
  {
131
  "epoch": 0.7623318385650224,
132
- "grad_norm": 0.25898587703704834,
133
  "learning_rate": 3.21490588442868e-05,
134
- "loss": 1.6834,
135
  "step": 85
136
  },
137
  {
138
  "epoch": 0.8071748878923767,
139
- "grad_norm": 0.2714327871799469,
140
  "learning_rate": 2.139469052572127e-05,
141
- "loss": 1.6902,
142
  "step": 90
143
  },
144
  {
145
  "epoch": 0.852017937219731,
146
- "grad_norm": 0.22857919335365295,
147
  "learning_rate": 1.2615062293021507e-05,
148
- "loss": 1.6917,
149
  "step": 95
150
  },
151
  {
152
  "epoch": 0.8968609865470852,
153
- "grad_norm": 0.2975883185863495,
154
  "learning_rate": 6.030737921409169e-06,
155
- "loss": 1.6887,
156
  "step": 100
157
  },
158
  {
159
  "epoch": 0.9417040358744395,
160
- "grad_norm": 0.2528745234012604,
161
  "learning_rate": 1.8071302737293295e-06,
162
- "loss": 1.6853,
163
  "step": 105
164
  },
165
  {
166
  "epoch": 0.9865470852017937,
167
- "grad_norm": 0.2566002309322357,
168
  "learning_rate": 5.0345761681491746e-08,
169
- "loss": 1.6835,
170
  "step": 110
171
  },
172
  {
173
  "epoch": 0.9955156950672646,
174
- "eval_loss": 2.371652841567993,
175
- "eval_runtime": 1.116,
176
- "eval_samples_per_second": 18.816,
177
- "eval_steps_per_second": 0.896,
178
  "step": 111
179
  },
180
  {
181
  "epoch": 0.9955156950672646,
182
  "step": 111,
183
  "total_flos": 6.554327132594504e+17,
184
- "train_loss": 1.8387699009061933,
185
- "train_runtime": 1914.8973,
186
- "train_samples_per_second": 14.857,
187
- "train_steps_per_second": 0.058
188
  }
189
  ],
190
  "logging_steps": 5,
 
10
  "log_history": [
11
  {
12
  "epoch": 0.008968609865470852,
13
+ "grad_norm": 5.0164899826049805,
14
  "learning_rate": 1.6666666666666667e-05,
15
+ "loss": 2.5202,
16
  "step": 1
17
  },
18
  {
19
  "epoch": 0.04484304932735426,
20
+ "grad_norm": 3.011996269226074,
21
  "learning_rate": 8.333333333333334e-05,
22
+ "loss": 2.5232,
23
  "step": 5
24
  },
25
  {
26
  "epoch": 0.08968609865470852,
27
+ "grad_norm": 0.789746880531311,
28
  "learning_rate": 0.0001666666666666667,
29
+ "loss": 2.4146,
30
  "step": 10
31
  },
32
  {
33
  "epoch": 0.13452914798206278,
34
+ "grad_norm": 0.8607348799705505,
35
  "learning_rate": 0.00019954719225730847,
36
+ "loss": 2.2464,
37
  "step": 15
38
  },
39
  {
40
  "epoch": 0.17937219730941703,
41
+ "grad_norm": 0.6927841305732727,
42
  "learning_rate": 0.00019679487013963564,
43
+ "loss": 2.0564,
44
  "step": 20
45
  },
46
  {
47
  "epoch": 0.2242152466367713,
48
+ "grad_norm": 0.569992184638977,
49
  "learning_rate": 0.00019161084574320696,
50
+ "loss": 1.9377,
51
  "step": 25
52
  },
53
  {
54
  "epoch": 0.26905829596412556,
55
+ "grad_norm": 0.3795563876628876,
56
  "learning_rate": 0.00018412535328311814,
57
+ "loss": 1.8624,
58
  "step": 30
59
  },
60
  {
61
  "epoch": 0.31390134529147984,
62
+ "grad_norm": 0.2764267325401306,
63
  "learning_rate": 0.0001745264449675755,
64
+ "loss": 1.8067,
65
  "step": 35
66
  },
67
  {
68
  "epoch": 0.35874439461883406,
69
+ "grad_norm": 0.2873355448246002,
70
  "learning_rate": 0.00016305526670845226,
71
+ "loss": 1.7802,
72
  "step": 40
73
  },
74
  {
75
  "epoch": 0.40358744394618834,
76
+ "grad_norm": 0.2722155749797821,
77
  "learning_rate": 0.00015000000000000001,
78
+ "loss": 1.7499,
79
  "step": 45
80
  },
81
  {
82
  "epoch": 0.4484304932735426,
83
+ "grad_norm": 0.24966852366924286,
84
  "learning_rate": 0.00013568862215918717,
85
+ "loss": 1.7376,
86
  "step": 50
87
  },
88
  {
89
  "epoch": 0.49327354260089684,
90
+ "grad_norm": 0.23957866430282593,
91
  "learning_rate": 0.00012048066680651908,
92
+ "loss": 1.7284,
93
  "step": 55
94
  },
95
  {
96
  "epoch": 0.5381165919282511,
97
+ "grad_norm": 0.2508368492126465,
98
  "learning_rate": 0.00010475819158237425,
99
+ "loss": 1.7107,
100
  "step": 60
101
  },
102
  {
103
  "epoch": 0.5829596412556054,
104
+ "grad_norm": 0.24922572076320648,
105
  "learning_rate": 8.891618000989891e-05,
106
+ "loss": 1.6965,
107
  "step": 65
108
  },
109
  {
110
  "epoch": 0.6278026905829597,
111
+ "grad_norm": 0.23529289662837982,
112
  "learning_rate": 7.335261863099651e-05,
113
+ "loss": 1.6994,
114
  "step": 70
115
  },
116
  {
117
  "epoch": 0.672645739910314,
118
+ "grad_norm": 0.24354910850524902,
119
  "learning_rate": 5.845849869981137e-05,
120
+ "loss": 1.6859,
121
  "step": 75
122
  },
123
  {
124
  "epoch": 0.7174887892376681,
125
+ "grad_norm": 0.24570442736148834,
126
  "learning_rate": 4.4607993613388976e-05,
127
+ "loss": 1.6771,
128
  "step": 80
129
  },
130
  {
131
  "epoch": 0.7623318385650224,
132
+ "grad_norm": 0.2550687789916992,
133
  "learning_rate": 3.21490588442868e-05,
134
+ "loss": 1.6749,
135
  "step": 85
136
  },
137
  {
138
  "epoch": 0.8071748878923767,
139
+ "grad_norm": 0.26918160915374756,
140
  "learning_rate": 2.139469052572127e-05,
141
+ "loss": 1.6817,
142
  "step": 90
143
  },
144
  {
145
  "epoch": 0.852017937219731,
146
+ "grad_norm": 0.2187623828649521,
147
  "learning_rate": 1.2615062293021507e-05,
148
+ "loss": 1.6831,
149
  "step": 95
150
  },
151
  {
152
  "epoch": 0.8968609865470852,
153
+ "grad_norm": 0.28687548637390137,
154
  "learning_rate": 6.030737921409169e-06,
155
+ "loss": 1.68,
156
  "step": 100
157
  },
158
  {
159
  "epoch": 0.9417040358744395,
160
+ "grad_norm": 0.2411273717880249,
161
  "learning_rate": 1.8071302737293295e-06,
162
+ "loss": 1.6766,
163
  "step": 105
164
  },
165
  {
166
  "epoch": 0.9865470852017937,
167
+ "grad_norm": 0.24854984879493713,
168
  "learning_rate": 5.0345761681491746e-08,
169
+ "loss": 1.6751,
170
  "step": 110
171
  },
172
  {
173
  "epoch": 0.9955156950672646,
174
+ "eval_loss": 2.3661022186279297,
175
+ "eval_runtime": 0.797,
176
+ "eval_samples_per_second": 26.35,
177
+ "eval_steps_per_second": 1.255,
178
  "step": 111
179
  },
180
  {
181
  "epoch": 0.9955156950672646,
182
  "step": 111,
183
  "total_flos": 6.554327132594504e+17,
184
+ "train_loss": 1.8339371971181921,
185
+ "train_runtime": 536.3339,
186
+ "train_samples_per_second": 53.043,
187
+ "train_steps_per_second": 0.207
188
  }
189
  ],
190
  "logging_steps": 5,