Edit model card

only_vary_model1

This model is a fine-tuned version of google/flan-t5-base on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 0.2400
  • Rouge1: 62.207
  • Rouge2: 48.2441
  • Rougel: 57.479
  • Rougelsum: 62.039
  • Gen Len: 16.3689

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-05
  • train_batch_size: 8
  • eval_batch_size: 16
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_steps: 200
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Rouge1 Rouge2 Rougel Rougelsum Gen Len
11.6904 0.7 200 0.5014 40.0947 28.9027 39.0539 40.0066 6.986
0.4064 1.4 400 0.2939 58.7275 43.2853 55.0201 58.5308 16.264
0.3125 2.1 600 0.2749 60.4914 45.7417 56.4753 60.3543 16.1503
0.2836 2.8 800 0.2576 60.3004 45.6356 55.9723 60.1254 16.2692
0.2634 3.5 1000 0.2519 61.3543 47.3391 57.2049 61.1618 16.3724
0.2608 4.2 1200 0.2522 61.538 46.8595 56.8033 61.3561 16.3409
0.2387 4.9 1400 0.2439 60.7253 46.2206 56.342 60.57 16.3357
0.2253 5.59 1600 0.2479 60.9499 46.7452 56.8289 60.8009 16.4178
0.2299 6.29 1800 0.2416 61.6976 47.212 57.1834 61.558 16.3077
0.2174 6.99 2000 0.2418 61.829 47.2977 57.3766 61.718 16.3112
0.2089 7.69 2200 0.2389 62.2459 47.6275 57.1022 62.1048 16.299
0.2127 8.39 2400 0.2402 61.8535 47.7161 57.2504 61.7496 16.4196
0.2097 9.09 2600 0.2402 62.242 47.6438 57.0938 62.0981 16.299
0.2036 9.79 2800 0.2400 62.207 48.2441 57.479 62.039 16.3689

Framework versions

  • Transformers 4.34.0
  • Pytorch 2.0.1+cu117
  • Datasets 2.14.4
  • Tokenizers 0.14.0
Downloads last month
0
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for tanvirsrbd1/only_vary_model1

Finetuned
(628)
this model