bert-base-uncased-BiLSTM-Optiparam-ADVQA36K-V10-frozen

This model is a fine-tuned version of allistair99/bert-base-uncased-ADVQA36K-V1 on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 3.8537

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 3e-05
  • train_batch_size: 6
  • eval_batch_size: 60
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 3

Training results

Training Loss Epoch Step Validation Loss
2.5455 0.0599 100 3.2561
0.8163 0.1198 200 3.6960
0.6833 0.1796 300 3.7703
0.6589 0.2395 400 3.6731
0.6259 0.2994 500 3.6294
0.6005 0.3593 600 3.8198
0.6102 0.4192 700 3.6329
0.5612 0.4790 800 3.9340
0.5562 0.5389 900 3.8636
0.6321 0.5988 1000 3.8534
0.6266 0.6587 1100 3.8581
0.5957 0.7186 1200 3.6847
0.6365 0.7784 1300 3.6420
0.5981 0.8383 1400 3.8763
0.6026 0.8982 1500 3.8009
0.6346 0.9581 1600 3.5014
0.5852 1.0180 1700 3.8213
0.6028 1.0778 1800 3.9848
0.5332 1.1377 1900 4.1392
0.5743 1.1976 2000 3.9228
0.5154 1.2575 2100 3.9906
0.5599 1.3174 2200 3.8916
0.5359 1.3772 2300 3.9310
0.5824 1.4371 2400 3.8373
0.5521 1.4970 2500 4.0342
0.5997 1.5569 2600 3.9670
0.5812 1.6168 2700 3.9608
0.5845 1.6766 2800 3.8695
0.638 1.7365 2900 3.6859
0.5949 1.7964 3000 3.8397
0.5571 1.8563 3100 3.9110
0.6531 1.9162 3200 3.8701
0.6434 1.9760 3300 3.7682
0.5097 2.0359 3400 4.0508
0.4267 2.0958 3500 4.2902
0.413 2.1557 3600 4.3331
0.4525 2.2156 3700 4.2495
0.5343 2.2754 3800 4.1981
0.5311 2.3353 3900 4.2015
0.5436 2.3952 4000 4.1445
0.5387 2.4551 4100 4.0878
0.5022 2.5150 4200 4.1251
0.5622 2.5749 4300 4.0905
0.5751 2.6347 4400 4.0135
0.544 2.6946 4500 4.0193
0.6347 2.7545 4600 3.9842
0.694 2.8144 4700 3.9273
0.6522 2.8743 4800 3.8841
0.7517 2.9341 4900 3.8684
0.8092 2.9940 5000 3.8537

Framework versions

  • Transformers 4.41.0
  • Pytorch 2.6.0+cu124
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
12
Safetensors
Model size
112M params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for allistair99/bert-base-uncased-BiLSTM-Optiparam-ADVQA36K-V10-frozen