turkish-hs-group-prediction
This model is a fine-tuned version of dbmdz/bert-base-turkish-uncased on the None dataset. It achieves the following results on the evaluation set:
- Loss: 0.1538
- Accuracy: 0.9742
- Macro F1: 0.9643
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-06
- train_batch_size: 16
- eval_batch_size: 20
- seed: 42
- optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: linear
- num_epochs: 10
Training results
Training Loss | Epoch | Step | Validation Loss | Accuracy | Macro F1 |
---|---|---|---|---|---|
2.0916 | 0.1391 | 100 | 1.7790 | 0.4358 | 0.1744 |
1.5505 | 0.2782 | 200 | 1.1988 | 0.6995 | 0.4707 |
0.9851 | 0.4172 | 300 | 0.6415 | 0.8944 | 0.7400 |
0.5924 | 0.5563 | 400 | 0.3818 | 0.9139 | 0.7976 |
0.3964 | 0.6954 | 500 | 0.2921 | 0.9358 | 0.8834 |
0.2746 | 0.8345 | 600 | 0.2091 | 0.9538 | 0.9189 |
0.1811 | 0.9736 | 700 | 0.1834 | 0.9570 | 0.9298 |
0.1676 | 1.1127 | 800 | 0.1613 | 0.9593 | 0.9318 |
0.1247 | 1.2517 | 900 | 0.1538 | 0.9609 | 0.9410 |
0.1234 | 1.3908 | 1000 | 0.1422 | 0.9656 | 0.9466 |
0.1134 | 1.5299 | 1100 | 0.1286 | 0.9679 | 0.9507 |
0.1162 | 1.6690 | 1200 | 0.1442 | 0.9679 | 0.9538 |
0.072 | 1.8081 | 1300 | 0.1394 | 0.9687 | 0.9508 |
0.1049 | 1.9471 | 1400 | 0.1380 | 0.9687 | 0.9499 |
0.0757 | 2.0862 | 1500 | 0.1409 | 0.9671 | 0.9483 |
0.0777 | 2.2253 | 1600 | 0.1186 | 0.9679 | 0.9505 |
0.0745 | 2.3644 | 1700 | 0.1182 | 0.9726 | 0.9608 |
0.0676 | 2.5035 | 1800 | 0.1368 | 0.9679 | 0.9547 |
0.0692 | 2.6426 | 1900 | 0.1250 | 0.9687 | 0.9547 |
0.076 | 2.7816 | 2000 | 0.1264 | 0.9726 | 0.9621 |
0.0704 | 2.9207 | 2100 | 0.1361 | 0.9703 | 0.9601 |
0.0427 | 3.0598 | 2200 | 0.1362 | 0.9703 | 0.9605 |
0.0456 | 3.1989 | 2300 | 0.1258 | 0.9734 | 0.9638 |
0.0584 | 3.3380 | 2400 | 0.1306 | 0.9710 | 0.9610 |
0.049 | 3.4771 | 2500 | 0.1258 | 0.9726 | 0.9631 |
0.0446 | 3.6161 | 2600 | 0.1294 | 0.9734 | 0.9642 |
0.0688 | 3.7552 | 2700 | 0.1473 | 0.9718 | 0.9620 |
0.0393 | 3.8943 | 2800 | 0.1335 | 0.9734 | 0.9642 |
0.0447 | 4.0334 | 2900 | 0.1456 | 0.9710 | 0.9616 |
0.0453 | 4.1725 | 3000 | 0.1373 | 0.9726 | 0.9633 |
0.0313 | 4.3115 | 3100 | 0.1403 | 0.9742 | 0.9647 |
0.0396 | 4.4506 | 3200 | 0.1393 | 0.9734 | 0.9636 |
0.0425 | 4.5897 | 3300 | 0.1296 | 0.9726 | 0.9623 |
0.0481 | 4.7288 | 3400 | 0.1333 | 0.9726 | 0.9631 |
0.0358 | 4.8679 | 3500 | 0.1456 | 0.9742 | 0.9648 |
0.0228 | 5.0070 | 3600 | 0.1482 | 0.9734 | 0.9633 |
0.029 | 5.1460 | 3700 | 0.1424 | 0.9750 | 0.9650 |
0.0237 | 5.2851 | 3800 | 0.1513 | 0.9726 | 0.9625 |
0.0235 | 5.4242 | 3900 | 0.1504 | 0.9734 | 0.9641 |
0.0471 | 5.5633 | 4000 | 0.1553 | 0.9726 | 0.9626 |
0.029 | 5.7024 | 4100 | 0.1343 | 0.9734 | 0.9635 |
0.0311 | 5.8414 | 4200 | 0.1331 | 0.9734 | 0.9635 |
0.0187 | 5.9805 | 4300 | 0.1495 | 0.9742 | 0.9645 |
0.0228 | 6.1196 | 4400 | 0.1549 | 0.9734 | 0.9637 |
0.0246 | 6.2587 | 4500 | 0.1510 | 0.9734 | 0.9637 |
0.0243 | 6.3978 | 4600 | 0.1378 | 0.9742 | 0.9645 |
0.0166 | 6.5369 | 4700 | 0.1485 | 0.9726 | 0.9629 |
0.0217 | 6.6759 | 4800 | 0.1519 | 0.9734 | 0.9637 |
0.0304 | 6.8150 | 4900 | 0.1436 | 0.9742 | 0.9645 |
0.0322 | 6.9541 | 5000 | 0.1434 | 0.9726 | 0.9618 |
0.0209 | 7.0932 | 5100 | 0.1475 | 0.9742 | 0.9643 |
0.0226 | 7.2323 | 5200 | 0.1380 | 0.9750 | 0.9654 |
0.0081 | 7.3713 | 5300 | 0.1422 | 0.9742 | 0.9643 |
0.0104 | 7.5104 | 5400 | 0.1452 | 0.9742 | 0.9641 |
0.0184 | 7.6495 | 5500 | 0.1438 | 0.9734 | 0.9638 |
0.0196 | 7.7886 | 5600 | 0.1449 | 0.9742 | 0.9647 |
0.0327 | 7.9277 | 5700 | 0.1480 | 0.9734 | 0.9635 |
0.0172 | 8.0668 | 5800 | 0.1444 | 0.9742 | 0.9647 |
0.0132 | 8.2058 | 5900 | 0.1480 | 0.9750 | 0.9655 |
0.0209 | 8.3449 | 6000 | 0.1446 | 0.9750 | 0.9655 |
0.0181 | 8.4840 | 6100 | 0.1454 | 0.9726 | 0.9619 |
0.0181 | 8.6231 | 6200 | 0.1458 | 0.9757 | 0.9659 |
0.0172 | 8.7622 | 6300 | 0.1505 | 0.9734 | 0.9631 |
0.0134 | 8.9013 | 6400 | 0.1616 | 0.9734 | 0.9634 |
0.0196 | 9.0403 | 6500 | 0.1607 | 0.9734 | 0.9634 |
0.007 | 9.1794 | 6600 | 0.1590 | 0.9734 | 0.9634 |
0.0206 | 9.3185 | 6700 | 0.1585 | 0.9734 | 0.9634 |
0.0161 | 9.4576 | 6800 | 0.1573 | 0.9734 | 0.9634 |
0.0134 | 9.5967 | 6900 | 0.1553 | 0.9734 | 0.9634 |
0.0158 | 9.7357 | 7000 | 0.1539 | 0.9734 | 0.9634 |
0.0109 | 9.8748 | 7100 | 0.1538 | 0.9742 | 0.9643 |
Framework versions
- Transformers 4.49.0
- Pytorch 2.5.1+cu124
- Datasets 3.3.2
- Tokenizers 0.21.0
- Downloads last month
- 196
Inference Providers
NEW
This model is not currently available via any of the supported Inference Providers.
Model tree for HrantDinkFoundation/turkish-hs-group-prediction
Base model
dbmdz/bert-base-turkish-uncased