gliner-bi-small-v1.0_2025-02-11_13-01-38

This model is a fine-tuned version of on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 39.4311

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.0005
  • train_batch_size: 4
  • eval_batch_size: 2
  • seed: 42
  • optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_ratio: 0.1
  • num_epochs: 4

Training results

Training Loss Epoch Step Validation Loss
31.572 0.1650 100 122.2061
20.4799 0.3300 200 158.5707
36.2449 0.4950 300 118.9915
19.9128 0.6601 400 98.7558
16.3263 0.8251 500 126.5938
15.1254 0.9901 600 124.2355
13.388 1.1551 700 111.7542
12.3434 1.3201 800 101.8177
12.3511 1.4851 900 74.6518
11.1384 1.6502 1000 70.6329
11.9784 1.8152 1100 72.1667
10.6829 1.9802 1200 68.7952
7.8355 2.1452 1300 72.6674
8.1642 2.3102 1400 60.0159
7.5026 2.4752 1500 66.2888
7.6514 2.6403 1600 86.6573
8.2808 2.8053 1700 57.6119
6.7184 2.9703 1800 56.0169
5.4226 3.1353 1900 44.2044
5.4276 3.3003 2000 54.2251
4.9535 3.4653 2100 42.1324
4.6889 3.6304 2200 43.5530
3.8583 3.7954 2300 38.7769
3.8563 3.9604 2400 39.4311

Framework versions

  • Transformers 4.48.2
  • Pytorch 2.5.1+cu124
  • Datasets 3.2.0
  • Tokenizers 0.21.0
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no pipeline_tag.