HuyTran1301 commited on
Commit
4aa85ba
·
verified ·
1 Parent(s): a47270d

End of training

Browse files
Files changed (4) hide show
  1. README.md +13 -10
  2. model.safetensors +1 -1
  3. training_args.bin +1 -1
  4. training_log_clean.csv +64 -5
README.md CHANGED
@@ -18,10 +18,10 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  This model is a fine-tuned version of [Salesforce/codet5-base](https://huggingface.co/Salesforce/codet5-base) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 0.9820
22
- - Rouge1: 0.0972
23
- - Rouge2: 0.0667
24
- - Rougel: 0.0972
25
 
26
  ## Model description
27
 
@@ -46,18 +46,21 @@ The following hyperparameters were used during training:
46
  - seed: 42
47
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
48
  - lr_scheduler_type: linear
49
- - num_epochs: 5
50
  - mixed_precision_training: Native AMP
51
 
52
  ### Training results
53
 
54
  | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel |
55
  |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|
56
- | 5.0107 | 1.0 | 1 | 6.5380 | 0.1056 | 0.0 | 0.1056 |
57
- | 5.0799 | 2.0 | 2 | 6.5380 | 0.1056 | 0.0 | 0.1056 |
58
- | 5.2912 | 3.0 | 3 | 1.1345 | 0.1472 | 0.0667 | 0.1472 |
59
- | 2.3361 | 4.0 | 4 | 1.0165 | 0.0972 | 0.0667 | 0.0972 |
60
- | 1.5034 | 5.0 | 5 | 0.9820 | 0.0972 | 0.0667 | 0.0972 |
 
 
 
61
 
62
 
63
  ### Framework versions
 
18
 
19
  This model is a fine-tuned version of [Salesforce/codet5-base](https://huggingface.co/Salesforce/codet5-base) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.5395
22
+ - Rouge1: 0.3357
23
+ - Rouge2: 0.1004
24
+ - Rougel: 0.3230
25
 
26
  ## Model description
27
 
 
46
  - seed: 42
47
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
48
  - lr_scheduler_type: linear
49
+ - num_epochs: 8
50
  - mixed_precision_training: Native AMP
51
 
52
  ### Training results
53
 
54
  | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel |
55
  |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|
56
+ | 1.2932 | 1.0 | 8 | 0.8983 | 0.2450 | 0.0317 | 0.2338 |
57
+ | 0.7191 | 2.0 | 16 | 0.7307 | 0.3123 | 0.0742 | 0.3011 |
58
+ | 0.812 | 3.0 | 24 | 0.6878 | 0.3304 | 0.0994 | 0.3186 |
59
+ | 0.4139 | 4.0 | 32 | 0.6456 | 0.3431 | 0.1057 | 0.3313 |
60
+ | 0.482 | 5.0 | 40 | 0.6134 | 0.3445 | 0.0942 | 0.3327 |
61
+ | 0.3675 | 6.0 | 48 | 0.5809 | 0.3318 | 0.0924 | 0.3210 |
62
+ | 0.4833 | 7.0 | 56 | 0.5530 | 0.3408 | 0.0963 | 0.3282 |
63
+ | 0.3695 | 8.0 | 64 | 0.5395 | 0.3357 | 0.1004 | 0.3230 |
64
 
65
 
66
  ### Framework versions
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f30bd7514a89ece80588fd76f09ab4fb9be1b5fb811c3455e4ae86c2fbac966d
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c64a1c6253971217ecd1d616dfba97266a64b5c4c10c00a0952956b987a9680
3
  size 891558696
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5ea26a31bdb1ee3ff20c3a1c14fa0a2b5def5a8faa992686b9bdc97baf752ccc
3
  size 5496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de5e70306183d11eac66f19ab5daed468226b186158b8c37315a7f2120150708
3
  size 5496
training_log_clean.csv CHANGED
@@ -1,6 +1,65 @@
1
  step,epoch,train_loss,eval_loss,rouge1,rouge2,rougeL
2
- 1,1.0,5.0107,6.538,0.1056,0.0,0.1056
3
- 2,2.0,5.0799,6.538,0.1056,0.0,0.1056
4
- 3,3.0,5.2912,1.1345,0.1472,0.0667,0.1472
5
- 4,4.0,2.3361,1.0165,0.0972,0.0667,0.0972
6
- 5,5.0,1.5034,0.982,0.0972,0.0667,0.0972
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  step,epoch,train_loss,eval_loss,rouge1,rouge2,rougeL
2
+ 1,0.125,4.3431,,,,
3
+ 2,0.25,4.4577,,,,
4
+ 3,0.375,4.256,,,,
5
+ 4,0.5,2.058,,,,
6
+ 5,0.625,1.1859,,,,
7
+ 6,0.75,1.1357,,,,
8
+ 7,0.875,1.1206,,,,
9
+ 8,1.0,1.2932,0.8983,0.245,0.0317,0.2338
10
+ 9,1.125,0.8722,,,,
11
+ 10,1.25,0.7869,,,,
12
+ 11,1.375,0.8525,,,,
13
+ 12,1.5,0.9085,,,,
14
+ 13,1.625,0.7161,,,,
15
+ 14,1.75,0.6706,,,,
16
+ 15,1.875,0.6845,,,,
17
+ 16,2.0,0.7191,0.7307,0.3123,0.0742,0.3011
18
+ 17,2.125,0.6398,,,,
19
+ 18,2.25,0.633,,,,
20
+ 19,2.375,0.6569,,,,
21
+ 20,2.5,0.6283,,,,
22
+ 21,2.625,0.6226,,,,
23
+ 22,2.75,0.5941,,,,
24
+ 23,2.875,0.7617,,,,
25
+ 24,3.0,0.812,0.6878,0.3304,0.0994,0.3186
26
+ 25,3.125,0.6705,,,,
27
+ 26,3.25,0.6552,,,,
28
+ 27,3.375,0.5242,,,,
29
+ 28,3.5,0.5521,,,,
30
+ 29,3.625,0.5372,,,,
31
+ 30,3.75,0.5754,,,,
32
+ 31,3.875,0.4648,,,,
33
+ 32,4.0,0.4139,0.6456,0.3431,0.1057,0.3313
34
+ 33,4.125,0.5441,,,,
35
+ 34,4.25,0.5423,,,,
36
+ 35,4.375,0.5342,,,,
37
+ 36,4.5,0.4525,,,,
38
+ 37,4.625,0.4669,,,,
39
+ 38,4.75,0.4311,,,,
40
+ 39,4.875,0.5024,,,,
41
+ 40,5.0,0.482,0.6134,0.3445,0.0942,0.3327
42
+ 41,5.125,0.4532,,,,
43
+ 42,5.25,0.417,,,,
44
+ 43,5.375,0.4782,,,,
45
+ 44,5.5,0.4652,,,,
46
+ 45,5.625,0.3913,,,,
47
+ 46,5.75,0.377,,,,
48
+ 47,5.875,0.4338,,,,
49
+ 48,6.0,0.3675,0.5809,0.3318,0.0924,0.321
50
+ 49,6.125,0.3642,,,,
51
+ 50,6.25,0.4242,,,,
52
+ 51,6.375,0.3509,,,,
53
+ 52,6.5,0.3426,,,,
54
+ 53,6.625,0.3705,,,,
55
+ 54,6.75,0.3472,,,,
56
+ 55,6.875,0.3631,,,,
57
+ 56,7.0,0.4833,0.553,0.3408,0.0963,0.3282
58
+ 57,7.125,0.3045,,,,
59
+ 58,7.25,0.3421,,,,
60
+ 59,7.375,0.3454,,,,
61
+ 60,7.5,0.396,,,,
62
+ 61,7.625,0.2711,,,,
63
+ 62,7.75,0.3038,,,,
64
+ 63,7.875,0.3511,,,,
65
+ 64,8.0,0.3695,0.5395,0.3357,0.1004,0.323