--- license: apache-2.0 tags: - simplification - generated_from_trainer metrics: - rouge model-index: - name: flan-t5-base-clara-med results: [] --- # flan-t5-base-clara-med This model is a fine-tuned version of [google/flan-t5-base](https://huggingface.co/google/flan-t5-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.2902 - Rouge1: 28.2733 - Rouge2: 15.323 - Rougel: 26.1421 - Rougelsum: 26.1589 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5.6e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | |:-------------:|:-----:|:-----:|:---------------:|:-------:|:-------:|:-------:|:---------:| | No log | 1.0 | 380 | 1.4639 | 26.4167 | 14.2257 | 24.5659 | 24.5948 | | No log | 2.0 | 760 | 1.3946 | 26.7094 | 14.6358 | 25.0516 | 25.075 | | 1.6466 | 3.0 | 1140 | 1.3480 | 27.3758 | 14.6821 | 25.5935 | 25.6007 | | 1.6466 | 4.0 | 1520 | 1.3221 | 28.0769 | 14.9721 | 26.131 | 26.1506 | | 1.3671 | 5.0 | 1900 | 1.2988 | 27.8019 | 14.9244 | 25.8242 | 25.8322 | | 1.3671 | 6.0 | 2280 | 1.2965 | 27.9071 | 15.3235 | 26.1385 | 26.104 | | 1.3671 | 7.0 | 2660 | 1.2802 | 28.1866 | 15.4793 | 26.301 | 26.3031 | | 1.2248 | 8.0 | 3040 | 1.2733 | 27.9974 | 15.4379 | 26.1087 | 26.1159 | | 1.2248 | 9.0 | 3420 | 1.2591 | 28.2545 | 15.5006 | 26.2812 | 26.3306 | | 1.1155 | 10.0 | 3800 | 1.2609 | 27.8029 | 15.0837 | 25.7989 | 25.8486 | | 1.1155 | 11.0 | 4180 | 1.2612 | 27.676 | 15.0786 | 25.6261 | 25.6458 | | 1.1155 | 12.0 | 4560 | 1.2616 | 27.6811 | 15.0935 | 25.6905 | 25.7125 | | 1.0337 | 13.0 | 4940 | 1.2562 | 27.88 | 15.2395 | 25.8875 | 25.8988 | | 1.0337 | 14.0 | 5320 | 1.2624 | 27.9858 | 15.2151 | 25.9785 | 26.0226 | | 0.9784 | 15.0 | 5700 | 1.2674 | 28.044 | 15.1312 | 25.8866 | 25.9514 | | 0.9784 | 16.0 | 6080 | 1.2588 | 28.1022 | 15.3599 | 26.0641 | 26.0762 | | 0.9784 | 17.0 | 6460 | 1.2676 | 27.864 | 15.1432 | 25.8981 | 25.9221 | | 0.9246 | 18.0 | 6840 | 1.2620 | 27.8826 | 15.1457 | 25.8041 | 25.8971 | | 0.9246 | 19.0 | 7220 | 1.2671 | 27.965 | 15.0059 | 25.94 | 25.9831 | | 0.8891 | 20.0 | 7600 | 1.2733 | 28.3035 | 15.3041 | 26.2411 | 26.2723 | | 0.8891 | 21.0 | 7980 | 1.2748 | 28.5205 | 15.4851 | 26.4543 | 26.4725 | | 0.8891 | 22.0 | 8360 | 1.2793 | 28.3018 | 15.3251 | 26.2781 | 26.3203 | | 0.8578 | 23.0 | 8740 | 1.2788 | 28.039 | 15.238 | 25.9371 | 25.9856 | | 0.8578 | 24.0 | 9120 | 1.2901 | 28.3312 | 15.3396 | 26.1722 | 26.1993 | | 0.8299 | 25.0 | 9500 | 1.2863 | 28.0727 | 15.0182 | 25.91 | 25.9577 | | 0.8299 | 26.0 | 9880 | 1.2845 | 28.1828 | 15.1338 | 26.039 | 26.0493 | | 0.8299 | 27.0 | 10260 | 1.2819 | 28.1547 | 15.091 | 26.0256 | 26.0346 | | 0.8137 | 28.0 | 10640 | 1.2859 | 28.2203 | 15.3225 | 26.1493 | 26.1591 | | 0.8137 | 29.0 | 11020 | 1.2902 | 28.2459 | 15.3142 | 26.1283 | 26.1382 | | 0.8061 | 30.0 | 11400 | 1.2902 | 28.2733 | 15.323 | 26.1421 | 26.1589 | ### Framework versions - Transformers 4.25.1 - Pytorch 1.13.0 - Datasets 2.8.0 - Tokenizers 0.12.1