multilingual-e5-base-edu-scorer-lr3e4-bs32

This model is a fine-tuned version of intfloat/multilingual-e5-base on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7551
  • Precision: 0.3916
  • Recall: 0.3720
  • F1 Macro: 0.3704
  • Accuracy: 0.4709

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.0003
  • train_batch_size: 64
  • eval_batch_size: 64
  • seed: 42
  • optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: linear
  • num_epochs: 20
  • mixed_precision_training: Native AMP

Training results

Training Loss Epoch Step Validation Loss Precision Recall F1 Macro Accuracy
No log 0 0 4.8374 0.0513 0.1667 0.0785 0.3080
1.0429 0.6793 1000 0.9472 0.3859 0.3207 0.3148 0.4320
1.0195 1.3587 2000 0.8833 0.4146 0.3355 0.3344 0.4225
1.0332 2.0380 3000 0.8693 0.4213 0.3342 0.3294 0.4111
0.9677 2.7174 4000 0.8591 0.4387 0.3494 0.3501 0.4213
0.9779 3.3967 5000 0.8598 0.4207 0.3584 0.3632 0.4550
0.932 4.0761 6000 0.8469 0.4429 0.3681 0.3717 0.4496
0.9421 4.7554 7000 0.8427 0.4376 0.3658 0.3717 0.4655
0.9552 5.4348 8000 0.8377 0.4364 0.3610 0.3682 0.4607
0.9298 6.1141 9000 0.8297 0.4580 0.3841 0.3918 0.4655
0.9307 6.7935 10000 0.8211 0.4542 0.3728 0.3786 0.4582
0.8988 7.4728 11000 0.8164 0.4440 0.3584 0.3597 0.4480
0.9219 8.1522 12000 0.8119 0.4617 0.3704 0.3763 0.4534
0.9289 8.8315 13000 0.8037 0.4422 0.3710 0.3780 0.4706
0.9109 9.5109 14000 0.8214 0.4588 0.3643 0.3659 0.4363
0.9017 10.1902 15000 0.8189 0.4425 0.3839 0.3864 0.4534
0.9117 10.8696 16000 0.8061 0.4542 0.3778 0.3837 0.4595
0.8836 11.5489 17000 0.7869 0.4671 0.3753 0.3825 0.4730
0.8749 12.2283 18000 0.7920 0.4604 0.3793 0.3863 0.4665
0.9009 12.9076 19000 0.7990 0.4645 0.3689 0.3753 0.4786
0.8984 13.5870 20000 0.7838 0.4652 0.3806 0.3899 0.4754
0.8525 14.2663 21000 0.7947 0.4478 0.3693 0.3758 0.4835
0.8514 14.9457 22000 0.7822 0.4734 0.3882 0.3976 0.4720
0.8796 15.625 23000 0.7804 0.4810 0.3823 0.3917 0.4706
0.8722 16.3043 24000 0.7820 0.4803 0.3819 0.3920 0.4738
0.8712 16.9837 25000 0.7815 0.4716 0.3845 0.3939 0.4679
0.8824 17.6630 26000 0.7827 0.4680 0.3726 0.3793 0.4792
0.8344 18.3424 27000 0.7792 0.4653 0.3760 0.3833 0.4810
0.8291 19.0217 28000 0.7755 0.4636 0.3806 0.3889 0.4774
0.8287 19.7011 29000 0.7754 0.4759 0.3861 0.3951 0.4762

Framework versions

  • Transformers 4.55.0
  • Pytorch 2.6.0.dev20241112+cu121
  • Datasets 4.0.0
  • Tokenizers 0.21.4
Downloads last month
30
Safetensors
Model size
278M params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for AngelinaZanardi/multilingual-e5-base-edu-scorer-lr3e4-bs32-swe

Finetuned
(88)
this model