End of training
Browse files- README.md +13 -10
- model.safetensors +1 -1
- training_args.bin +1 -1
- training_log_clean.csv +64 -5
README.md
CHANGED
@@ -18,10 +18,10 @@ should probably proofread and complete it, then remove this comment. -->
|
|
18 |
|
19 |
This model is a fine-tuned version of [Salesforce/codet5-base](https://huggingface.co/Salesforce/codet5-base) on the None dataset.
|
20 |
It achieves the following results on the evaluation set:
|
21 |
-
- Loss: 0.
|
22 |
-
- Rouge1: 0.
|
23 |
-
- Rouge2: 0.
|
24 |
-
- Rougel: 0.
|
25 |
|
26 |
## Model description
|
27 |
|
@@ -46,18 +46,21 @@ The following hyperparameters were used during training:
|
|
46 |
- seed: 42
|
47 |
- optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
|
48 |
- lr_scheduler_type: linear
|
49 |
-
- num_epochs:
|
50 |
- mixed_precision_training: Native AMP
|
51 |
|
52 |
### Training results
|
53 |
|
54 |
| Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel |
|
55 |
|:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|
|
56 |
-
|
|
57 |
-
|
|
58 |
-
|
|
59 |
-
|
|
60 |
-
|
|
|
|
|
|
|
|
61 |
|
62 |
|
63 |
### Framework versions
|
|
|
18 |
|
19 |
This model is a fine-tuned version of [Salesforce/codet5-base](https://huggingface.co/Salesforce/codet5-base) on the None dataset.
|
20 |
It achieves the following results on the evaluation set:
|
21 |
+
- Loss: 0.5395
|
22 |
+
- Rouge1: 0.3357
|
23 |
+
- Rouge2: 0.1004
|
24 |
+
- Rougel: 0.3230
|
25 |
|
26 |
## Model description
|
27 |
|
|
|
46 |
- seed: 42
|
47 |
- optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
|
48 |
- lr_scheduler_type: linear
|
49 |
+
- num_epochs: 8
|
50 |
- mixed_precision_training: Native AMP
|
51 |
|
52 |
### Training results
|
53 |
|
54 |
| Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel |
|
55 |
|:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|
|
56 |
+
| 1.2932 | 1.0 | 8 | 0.8983 | 0.2450 | 0.0317 | 0.2338 |
|
57 |
+
| 0.7191 | 2.0 | 16 | 0.7307 | 0.3123 | 0.0742 | 0.3011 |
|
58 |
+
| 0.812 | 3.0 | 24 | 0.6878 | 0.3304 | 0.0994 | 0.3186 |
|
59 |
+
| 0.4139 | 4.0 | 32 | 0.6456 | 0.3431 | 0.1057 | 0.3313 |
|
60 |
+
| 0.482 | 5.0 | 40 | 0.6134 | 0.3445 | 0.0942 | 0.3327 |
|
61 |
+
| 0.3675 | 6.0 | 48 | 0.5809 | 0.3318 | 0.0924 | 0.3210 |
|
62 |
+
| 0.4833 | 7.0 | 56 | 0.5530 | 0.3408 | 0.0963 | 0.3282 |
|
63 |
+
| 0.3695 | 8.0 | 64 | 0.5395 | 0.3357 | 0.1004 | 0.3230 |
|
64 |
|
65 |
|
66 |
### Framework versions
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 891558696
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9c64a1c6253971217ecd1d616dfba97266a64b5c4c10c00a0952956b987a9680
|
3 |
size 891558696
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 5496
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:de5e70306183d11eac66f19ab5daed468226b186158b8c37315a7f2120150708
|
3 |
size 5496
|
training_log_clean.csv
CHANGED
@@ -1,6 +1,65 @@
|
|
1 |
step,epoch,train_loss,eval_loss,rouge1,rouge2,rougeL
|
2 |
-
1,
|
3 |
-
2,
|
4 |
-
3,
|
5 |
-
4,
|
6 |
-
5,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
step,epoch,train_loss,eval_loss,rouge1,rouge2,rougeL
|
2 |
+
1,0.125,4.3431,,,,
|
3 |
+
2,0.25,4.4577,,,,
|
4 |
+
3,0.375,4.256,,,,
|
5 |
+
4,0.5,2.058,,,,
|
6 |
+
5,0.625,1.1859,,,,
|
7 |
+
6,0.75,1.1357,,,,
|
8 |
+
7,0.875,1.1206,,,,
|
9 |
+
8,1.0,1.2932,0.8983,0.245,0.0317,0.2338
|
10 |
+
9,1.125,0.8722,,,,
|
11 |
+
10,1.25,0.7869,,,,
|
12 |
+
11,1.375,0.8525,,,,
|
13 |
+
12,1.5,0.9085,,,,
|
14 |
+
13,1.625,0.7161,,,,
|
15 |
+
14,1.75,0.6706,,,,
|
16 |
+
15,1.875,0.6845,,,,
|
17 |
+
16,2.0,0.7191,0.7307,0.3123,0.0742,0.3011
|
18 |
+
17,2.125,0.6398,,,,
|
19 |
+
18,2.25,0.633,,,,
|
20 |
+
19,2.375,0.6569,,,,
|
21 |
+
20,2.5,0.6283,,,,
|
22 |
+
21,2.625,0.6226,,,,
|
23 |
+
22,2.75,0.5941,,,,
|
24 |
+
23,2.875,0.7617,,,,
|
25 |
+
24,3.0,0.812,0.6878,0.3304,0.0994,0.3186
|
26 |
+
25,3.125,0.6705,,,,
|
27 |
+
26,3.25,0.6552,,,,
|
28 |
+
27,3.375,0.5242,,,,
|
29 |
+
28,3.5,0.5521,,,,
|
30 |
+
29,3.625,0.5372,,,,
|
31 |
+
30,3.75,0.5754,,,,
|
32 |
+
31,3.875,0.4648,,,,
|
33 |
+
32,4.0,0.4139,0.6456,0.3431,0.1057,0.3313
|
34 |
+
33,4.125,0.5441,,,,
|
35 |
+
34,4.25,0.5423,,,,
|
36 |
+
35,4.375,0.5342,,,,
|
37 |
+
36,4.5,0.4525,,,,
|
38 |
+
37,4.625,0.4669,,,,
|
39 |
+
38,4.75,0.4311,,,,
|
40 |
+
39,4.875,0.5024,,,,
|
41 |
+
40,5.0,0.482,0.6134,0.3445,0.0942,0.3327
|
42 |
+
41,5.125,0.4532,,,,
|
43 |
+
42,5.25,0.417,,,,
|
44 |
+
43,5.375,0.4782,,,,
|
45 |
+
44,5.5,0.4652,,,,
|
46 |
+
45,5.625,0.3913,,,,
|
47 |
+
46,5.75,0.377,,,,
|
48 |
+
47,5.875,0.4338,,,,
|
49 |
+
48,6.0,0.3675,0.5809,0.3318,0.0924,0.321
|
50 |
+
49,6.125,0.3642,,,,
|
51 |
+
50,6.25,0.4242,,,,
|
52 |
+
51,6.375,0.3509,,,,
|
53 |
+
52,6.5,0.3426,,,,
|
54 |
+
53,6.625,0.3705,,,,
|
55 |
+
54,6.75,0.3472,,,,
|
56 |
+
55,6.875,0.3631,,,,
|
57 |
+
56,7.0,0.4833,0.553,0.3408,0.0963,0.3282
|
58 |
+
57,7.125,0.3045,,,,
|
59 |
+
58,7.25,0.3421,,,,
|
60 |
+
59,7.375,0.3454,,,,
|
61 |
+
60,7.5,0.396,,,,
|
62 |
+
61,7.625,0.2711,,,,
|
63 |
+
62,7.75,0.3038,,,,
|
64 |
+
63,7.875,0.3511,,,,
|
65 |
+
64,8.0,0.3695,0.5395,0.3357,0.1004,0.323
|