underscore2 commited on
Commit
95032a5
·
verified ·
1 Parent(s): 023e6ed

End of training

Browse files
README.md CHANGED
@@ -18,8 +18,8 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  This model is a fine-tuned version of [answerdotai/ModernBERT-large](https://huggingface.co/answerdotai/ModernBERT-large) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 0.4925
22
- - Accuracy: 0.8698
23
 
24
  ## Model description
25
 
@@ -38,27 +38,31 @@ More information needed
38
  ### Training hyperparameters
39
 
40
  The following hyperparameters were used during training:
41
- - learning_rate: 6.000000000000001e-05
42
- - train_batch_size: 4
43
  - eval_batch_size: 1
44
  - seed: 42
45
- - gradient_accumulation_steps: 3
46
  - total_train_batch_size: 12
47
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
48
  - lr_scheduler_type: constant
49
- - num_epochs: 2
50
 
51
  ### Training results
52
 
53
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
54
  |:-------------:|:------:|:----:|:---------------:|:--------:|
55
- | 1.3307 | 0.2757 | 50 | 0.8133 | 0.5547 |
56
- | 2.4211 | 0.5515 | 100 | 0.6010 | 0.6615 |
57
- | 1.9701 | 0.8272 | 150 | 0.5964 | 0.6901 |
58
- | 0.2215 | 1.0993 | 200 | 0.6405 | 0.7109 |
59
- | 0.1582 | 1.375 | 250 | 0.4628 | 0.7995 |
60
- | 0.1519 | 1.6507 | 300 | 0.4207 | 0.7995 |
61
- | 0.8939 | 1.9265 | 350 | 0.4925 | 0.8698 |
 
 
 
 
62
 
63
 
64
  ### Framework versions
 
18
 
19
  This model is a fine-tuned version of [answerdotai/ModernBERT-large](https://huggingface.co/answerdotai/ModernBERT-large) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 3.2279
22
+ - Accuracy: 0.7059
23
 
24
  ## Model description
25
 
 
38
  ### Training hyperparameters
39
 
40
  The following hyperparameters were used during training:
41
+ - learning_rate: 0.00016
42
+ - train_batch_size: 6
43
  - eval_batch_size: 1
44
  - seed: 42
45
+ - gradient_accumulation_steps: 2
46
  - total_train_batch_size: 12
47
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
48
  - lr_scheduler_type: constant
49
+ - training_steps: 550
50
 
51
  ### Training results
52
 
53
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
54
  |:-------------:|:------:|:----:|:---------------:|:--------:|
55
+ | 2.1915 | 0.2762 | 50 | 1.1417 | 0.2588 |
56
+ | 1.3494 | 0.5525 | 100 | 0.6381 | 0.7412 |
57
+ | 1.3588 | 0.8287 | 150 | 0.6750 | 0.5922 |
58
+ | 1.0859 | 1.1050 | 200 | 0.8046 | 0.7373 |
59
+ | 1.5857 | 1.3812 | 250 | 0.6222 | 0.6667 |
60
+ | 0.5529 | 1.6575 | 300 | 1.0381 | 0.4471 |
61
+ | 1.2417 | 1.9337 | 350 | 0.9943 | 0.6980 |
62
+ | 0.0183 | 2.2099 | 400 | 2.7391 | 0.6941 |
63
+ | 0.0112 | 2.4862 | 450 | 2.3648 | 0.6706 |
64
+ | 0.0004 | 2.7624 | 500 | 2.3759 | 0.6667 |
65
+ | 0.0007 | 3.0387 | 550 | 3.2279 | 0.7059 |
66
 
67
 
68
  ### Framework versions
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:87628d2b0de10e5d43c0b694902db6258e7054f038139cc6d7910ec479f64277
3
  size 1583351632
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c1407fcae343a868d7a62b691efc045ae41735f93cedd3e3d1cf727b73680fa
3
  size 1583351632
runs/May18_18-01-17_14d9f458975d/events.out.tfevents.1747591281.14d9f458975d.1682.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b88a903a6bf206eed27bef0bdbc16350467be66bb391ada076169a1d82b9e142
3
+ size 125150
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7a309a5f23b85f577a3f228afdb935295dd6749573e006ae3017b701a6c98319
3
  size 5304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93cf1bc4c7c7c31a9964722836d8d34df522aa3ceb58aca1b1dfcdf75fbb91d6
3
  size 5304