rtdetr-v2-r50-cppe5-finetune-2
This model is a fine-tuned version of PekingU/rtdetr_v2_r50vd on an unknown dataset. It achieves the following results on the evaluation set:
- Loss: 15.7596
- Map: 0.243
- Map 50: 0.4016
- Map 75: 0.2288
- Map Small: 0.1622
- Map Medium: 0.1679
- Map Large: 0.4047
- Mar 1: 0.2498
- Mar 10: 0.5329
- Mar 100: 0.6054
- Mar Small: 0.4114
- Mar Medium: 0.5602
- Mar Large: 0.781
- Map Coverall: 0.4109
- Mar 100 Coverall: 0.7821
- Map Face Shield: 0.1194
- Mar 100 Face Shield: 0.6471
- Map Gloves: 0.2443
- Mar 100 Gloves: 0.4508
- Map Goggles: 0.065
- Mar 100 Goggles: 0.4586
- Map Mask: 0.3754
- Mar 100 Mask: 0.6882
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 300
- num_epochs: 2
Training results
Training Loss | Epoch | Step | Validation Loss | Map | Map 50 | Map 75 | Map Small | Map Medium | Map Large | Mar 1 | Mar 10 | Mar 100 | Mar Small | Mar Medium | Mar Large | Map Coverall | Mar 100 Coverall | Map Face Shield | Mar 100 Face Shield | Map Gloves | Mar 100 Gloves | Map Goggles | Mar 100 Goggles | Map Mask | Mar 100 Mask |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
No log | 1.0 | 107 | 35.1046 | 0.0395 | 0.0911 | 0.0256 | 0.0015 | 0.0104 | 0.0475 | 0.0733 | 0.1851 | 0.2494 | 0.0421 | 0.1906 | 0.365 | 0.1668 | 0.5491 | 0.0207 | 0.2278 | 0.0068 | 0.1759 | 0.0007 | 0.0877 | 0.0024 | 0.2062 |
No log | 2.0 | 214 | 18.0639 | 0.1779 | 0.3283 | 0.1719 | 0.0679 | 0.1 | 0.2844 | 0.2053 | 0.4086 | 0.4734 | 0.2576 | 0.3816 | 0.7174 | 0.4958 | 0.7032 | 0.0485 | 0.4873 | 0.1107 | 0.3246 | 0.0257 | 0.3754 | 0.2086 | 0.4764 |
Framework versions
- Transformers 4.50.0.dev0
- Pytorch 2.6.0+cu124
- Datasets 3.4.1
- Tokenizers 0.21.1
- Downloads last month
- 14
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support
Model tree for lillythomas/rtdetr-v2-r50-cppe5-finetune-2
Base model
PekingU/rtdetr_v2_r50vd