maximo-t5-normalize / README.md
maxadmin's picture
End of training
86e8e83
|
raw
history blame
2.49 kB
metadata
license: apache-2.0
base_model: google/flan-t5-small
tags:
  - generated_from_trainer
metrics:
  - rouge
model-index:
  - name: maximo-t5-normalize
    results: []

maximo-t5-normalize

This model is a fine-tuned version of google/flan-t5-small on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0064
  • Rouge1: 48.6395
  • Rouge2: 26.1905
  • Rougel: 47.7211
  • Rougelsum: 48.1633
  • Gen Len: 10.5714

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.0005
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Rouge1 Rouge2 Rougel Rougelsum Gen Len
No log 1.0 8 2.5594 10.2165 0.0 9.9567 10.2165 11.0
No log 2.0 16 2.0407 25.7143 14.2857 25.7143 25.7143 6.0
No log 3.0 24 1.7822 46.0317 33.3333 46.0317 45.7143 11.7143
No log 4.0 32 1.5058 50.3401 32.8571 49.1156 49.932 11.0
No log 5.0 40 1.2425 48.6395 26.1905 47.7211 48.1633 11.7143
No log 6.0 48 1.1249 52.8912 33.3333 52.2109 52.2109 13.4286
No log 7.0 56 1.0713 41.7687 21.4286 40.9524 41.7687 9.7143
No log 8.0 64 1.0475 46.1224 21.4286 45.1701 45.4422 10.1429
No log 9.0 72 1.0154 43.8095 16.6667 42.7211 42.9932 10.7143
No log 10.0 80 1.0064 48.6395 26.1905 47.7211 48.1633 10.5714

Framework versions

  • Transformers 4.35.2
  • Pytorch 2.1.0+cu118
  • Datasets 2.15.0
  • Tokenizers 0.15.0