turkish-hs-group-prediction

This model is a fine-tuned version of dbmdz/bert-base-turkish-uncased on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.1538
  • Accuracy: 0.9742
  • Macro F1: 0.9643

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-06
  • train_batch_size: 16
  • eval_batch_size: 20
  • seed: 42
  • optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Accuracy Macro F1
2.0916 0.1391 100 1.7790 0.4358 0.1744
1.5505 0.2782 200 1.1988 0.6995 0.4707
0.9851 0.4172 300 0.6415 0.8944 0.7400
0.5924 0.5563 400 0.3818 0.9139 0.7976
0.3964 0.6954 500 0.2921 0.9358 0.8834
0.2746 0.8345 600 0.2091 0.9538 0.9189
0.1811 0.9736 700 0.1834 0.9570 0.9298
0.1676 1.1127 800 0.1613 0.9593 0.9318
0.1247 1.2517 900 0.1538 0.9609 0.9410
0.1234 1.3908 1000 0.1422 0.9656 0.9466
0.1134 1.5299 1100 0.1286 0.9679 0.9507
0.1162 1.6690 1200 0.1442 0.9679 0.9538
0.072 1.8081 1300 0.1394 0.9687 0.9508
0.1049 1.9471 1400 0.1380 0.9687 0.9499
0.0757 2.0862 1500 0.1409 0.9671 0.9483
0.0777 2.2253 1600 0.1186 0.9679 0.9505
0.0745 2.3644 1700 0.1182 0.9726 0.9608
0.0676 2.5035 1800 0.1368 0.9679 0.9547
0.0692 2.6426 1900 0.1250 0.9687 0.9547
0.076 2.7816 2000 0.1264 0.9726 0.9621
0.0704 2.9207 2100 0.1361 0.9703 0.9601
0.0427 3.0598 2200 0.1362 0.9703 0.9605
0.0456 3.1989 2300 0.1258 0.9734 0.9638
0.0584 3.3380 2400 0.1306 0.9710 0.9610
0.049 3.4771 2500 0.1258 0.9726 0.9631
0.0446 3.6161 2600 0.1294 0.9734 0.9642
0.0688 3.7552 2700 0.1473 0.9718 0.9620
0.0393 3.8943 2800 0.1335 0.9734 0.9642
0.0447 4.0334 2900 0.1456 0.9710 0.9616
0.0453 4.1725 3000 0.1373 0.9726 0.9633
0.0313 4.3115 3100 0.1403 0.9742 0.9647
0.0396 4.4506 3200 0.1393 0.9734 0.9636
0.0425 4.5897 3300 0.1296 0.9726 0.9623
0.0481 4.7288 3400 0.1333 0.9726 0.9631
0.0358 4.8679 3500 0.1456 0.9742 0.9648
0.0228 5.0070 3600 0.1482 0.9734 0.9633
0.029 5.1460 3700 0.1424 0.9750 0.9650
0.0237 5.2851 3800 0.1513 0.9726 0.9625
0.0235 5.4242 3900 0.1504 0.9734 0.9641
0.0471 5.5633 4000 0.1553 0.9726 0.9626
0.029 5.7024 4100 0.1343 0.9734 0.9635
0.0311 5.8414 4200 0.1331 0.9734 0.9635
0.0187 5.9805 4300 0.1495 0.9742 0.9645
0.0228 6.1196 4400 0.1549 0.9734 0.9637
0.0246 6.2587 4500 0.1510 0.9734 0.9637
0.0243 6.3978 4600 0.1378 0.9742 0.9645
0.0166 6.5369 4700 0.1485 0.9726 0.9629
0.0217 6.6759 4800 0.1519 0.9734 0.9637
0.0304 6.8150 4900 0.1436 0.9742 0.9645
0.0322 6.9541 5000 0.1434 0.9726 0.9618
0.0209 7.0932 5100 0.1475 0.9742 0.9643
0.0226 7.2323 5200 0.1380 0.9750 0.9654
0.0081 7.3713 5300 0.1422 0.9742 0.9643
0.0104 7.5104 5400 0.1452 0.9742 0.9641
0.0184 7.6495 5500 0.1438 0.9734 0.9638
0.0196 7.7886 5600 0.1449 0.9742 0.9647
0.0327 7.9277 5700 0.1480 0.9734 0.9635
0.0172 8.0668 5800 0.1444 0.9742 0.9647
0.0132 8.2058 5900 0.1480 0.9750 0.9655
0.0209 8.3449 6000 0.1446 0.9750 0.9655
0.0181 8.4840 6100 0.1454 0.9726 0.9619
0.0181 8.6231 6200 0.1458 0.9757 0.9659
0.0172 8.7622 6300 0.1505 0.9734 0.9631
0.0134 8.9013 6400 0.1616 0.9734 0.9634
0.0196 9.0403 6500 0.1607 0.9734 0.9634
0.007 9.1794 6600 0.1590 0.9734 0.9634
0.0206 9.3185 6700 0.1585 0.9734 0.9634
0.0161 9.4576 6800 0.1573 0.9734 0.9634
0.0134 9.5967 6900 0.1553 0.9734 0.9634
0.0158 9.7357 7000 0.1539 0.9734 0.9634
0.0109 9.8748 7100 0.1538 0.9742 0.9643

Framework versions

  • Transformers 4.49.0
  • Pytorch 2.5.1+cu124
  • Datasets 3.3.2
  • Tokenizers 0.21.0
Downloads last month
196
Safetensors
Model size
112M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for HrantDinkFoundation/turkish-hs-group-prediction

Finetuned
(20)
this model