samsum_42

This model is a fine-tuned version of google/t5-v1_1-small on the samsum dataset. It achieves the following results on the evaluation set:

  • Loss: 2.1130
  • Rouge1: 38.9743
  • Rouge2: 17.0483
  • Rougel: 32.8088
  • Rougelsum: 36.1047
  • Gen Len: 20.1785

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-05
  • train_batch_size: 16
  • eval_batch_size: 64
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 5

Training results

Training Loss Epoch Step Gen Len Validation Loss Rouge1 Rouge2 Rougel Rougelsum
No log 1.0 461 14.9658 2.3319 30.3301 13.4627 26.1544 28.1275
No log 2.0 922 17.4230 2.1737 35.9382 15.9281 30.5125 33.3475
3.9209 3.0 1383 19.5892 2.1288 37.9339 16.6056 31.9002 35.0597
3.9209 4.0 1844 20.1381 2.1201 38.625 16.8095 32.5753 35.825
2.9358 5.0 2305 20.1785 2.1130 38.9743 17.0483 32.8088 36.1047

Framework versions

  • Transformers 4.39.3
  • Pytorch 2.3.0+cu121
  • Datasets 2.19.0
  • Tokenizers 0.15.2
Downloads last month
11
Safetensors
Model size
77M params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for jialicheng/samsum_t5-small

Finetuned
(12)
this model

Evaluation results