rakhman-llm commited on
Commit
b3c591c
·
verified ·
1 Parent(s): 28247ff

Training in progress, step 23000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e0f4dd5c7f06f434e4d8cbac5754e2c65f64bd46ac822a7e90462688dcf30e65
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be1cf4c3d10cfbf74e0b37f9cbc786fae12e5bbc6e3706973bfac0175c00ecce
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a565ab0875e5f8e974ebb52445ef3770a07896c54aaad2beab4684116687c64d
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bca9f5a43e8c43cfe44e64364c4d417027de463f09d30955f0ca8d99d5517c8a
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e4371c359f3f61a0b206a73cb497201135751765261ece06b66d5005fa4a3d9e
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:29008ac370065a7f89ca5e36ad7037a88abffeace17ded8b029713d8007c00c5
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4c6c8e99b80f348613f137bd980faaac34ec4fa21cbf485269f4beb839740a21
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95f8453b710c139b3558b89069d0f1dbb29ca8e0859ba342fa8818ead3f6d844
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 1.5,
5
  "eval_steps": 500,
6
- "global_step": 22500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1590,6 +1590,41 @@
1590
  "learning_rate": 1.0003111111111113e-05,
1591
  "loss": 0.0564,
1592
  "step": 22500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1593
  }
1594
  ],
1595
  "logging_steps": 100,
@@ -1609,7 +1644,7 @@
1609
  "attributes": {}
1610
  }
1611
  },
1612
- "total_flos": 5.48062101504e+16,
1613
  "train_batch_size": 4,
1614
  "trial_name": null,
1615
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 1.5333333333333332,
5
  "eval_steps": 500,
6
+ "global_step": 23000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1590
  "learning_rate": 1.0003111111111113e-05,
1591
  "loss": 0.0564,
1592
  "step": 22500
1593
+ },
1594
+ {
1595
+ "epoch": 1.5066666666666668,
1596
+ "grad_norm": 0.24926000833511353,
1597
+ "learning_rate": 9.958666666666668e-06,
1598
+ "loss": 0.0562,
1599
+ "step": 22600
1600
+ },
1601
+ {
1602
+ "epoch": 1.5133333333333332,
1603
+ "grad_norm": 0.19163891673088074,
1604
+ "learning_rate": 9.914222222222224e-06,
1605
+ "loss": 0.0574,
1606
+ "step": 22700
1607
+ },
1608
+ {
1609
+ "epoch": 1.52,
1610
+ "grad_norm": 0.12077998369932175,
1611
+ "learning_rate": 9.869777777777779e-06,
1612
+ "loss": 0.0615,
1613
+ "step": 22800
1614
+ },
1615
+ {
1616
+ "epoch": 1.5266666666666666,
1617
+ "grad_norm": 0.23265130817890167,
1618
+ "learning_rate": 9.825333333333334e-06,
1619
+ "loss": 0.0624,
1620
+ "step": 22900
1621
+ },
1622
+ {
1623
+ "epoch": 1.5333333333333332,
1624
+ "grad_norm": 0.20572605729103088,
1625
+ "learning_rate": 9.78088888888889e-06,
1626
+ "loss": 0.0545,
1627
+ "step": 23000
1628
  }
1629
  ],
1630
  "logging_steps": 100,
 
1644
  "attributes": {}
1645
  }
1646
  },
1647
+ "total_flos": 5.602412593152e+16,
1648
  "train_batch_size": 4,
1649
  "trial_name": null,
1650
  "trial_params": null