speecht5_dhivehi_tts_v7_from_scratch

This model was trained from scratch on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.3932

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 1e-05
  • train_batch_size: 92
  • eval_batch_size: 16
  • seed: 42
  • optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: cosine_with_restarts
  • lr_scheduler_warmup_steps: 2000
  • training_steps: 70000
  • mixed_precision_training: Native AMP

Training results

Training Loss Epoch Step Validation Loss
0.4842 4.1667 1000 0.4482
0.485 8.3333 2000 0.4464
0.4862 12.5 3000 0.4476
0.482 16.6667 4000 0.4442
0.4832 20.8333 5000 0.4435
0.4803 25.0 6000 0.4433
0.4784 29.1667 7000 0.4421
0.4769 33.3333 8000 0.4416
0.4705 37.5 9000 0.4374
0.4711 41.6667 10000 0.4369
0.4705 45.8333 11000 0.4356
0.4694 50.0 12000 0.4331
0.4675 54.1667 13000 0.4327
0.4666 58.3333 14000 0.4325
0.4643 62.5 15000 0.4301
0.4608 66.6667 16000 0.4277
0.4585 70.8333 17000 0.4254
0.4592 75.0 18000 0.4255
0.4597 79.1667 19000 0.4232
0.4564 83.3333 20000 0.4247
0.4543 87.5 21000 0.4205
0.4528 91.6667 22000 0.4189
0.4507 95.8333 23000 0.4183
0.4489 100.0 24000 0.4167
0.4503 104.1667 25000 0.4168
0.447 108.3333 26000 0.4154
0.4421 112.5 27000 0.4150
0.4444 116.6667 28000 0.4149
0.4431 120.8333 29000 0.4116
0.4411 125.0 30000 0.4118
0.4431 129.1667 31000 0.4097
0.4376 133.3333 32000 0.4072
0.4343 137.5 33000 0.4091
0.4398 141.6667 34000 0.4061
0.4362 145.8333 35000 0.4063
0.4365 150.0 36000 0.4039
0.4363 154.1667 37000 0.4042
0.4341 158.3333 38000 0.4029
0.4302 162.5 39000 0.4030
0.4307 166.6667 40000 0.4015
0.4321 170.8333 41000 0.3996
0.4301 175.0 42000 0.4018
0.4305 179.1667 43000 0.4000
0.4287 183.3333 44000 0.4005
0.4325 187.5 45000 0.4001
0.4287 191.6667 46000 0.3981
0.4272 195.8333 47000 0.3972
0.4249 200.0 48000 0.3968
0.427 204.1667 49000 0.3970
0.4271 208.3333 50000 0.3967
0.4247 212.5 51000 0.3958
0.4233 216.6667 52000 0.3953
0.4245 220.8333 53000 0.3956
0.4251 225.0 54000 0.3940
0.4266 229.1667 55000 0.3951
0.4203 233.3333 56000 0.3940
0.4235 237.5 57000 0.3934
0.4276 241.6667 58000 0.3929
0.4215 245.8333 59000 0.3938
0.4285 250.0 60000 0.3936
0.425 254.1667 61000 0.3936
0.4237 258.3333 62000 0.3933
0.4243 262.5 63000 0.3932

Framework versions

  • Transformers 4.48.0.dev0
  • Pytorch 2.5.1+cu121
  • Datasets 3.2.0
  • Tokenizers 0.21.0
Downloads last month
62
Safetensors
Model size
144M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.