rakhman-llm commited on
Commit
29e065e
·
verified ·
1 Parent(s): bcb2a06

Training in progress, step 8000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e017ff86b8a295f11193c500291f5991ea24bf880a065787121b3fb732878a01
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a43ee6984f5f990cd4d1d7d439666b7dfa7b56da1d97ae91a2c6727d6281cb2d
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8013718d7b486d4bea38838926eb12f405577f302b56db31ac78355ece57c039
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f67eef48f1db3ec75565c0e239b37d2a1a24b92d26f253f611940b1cf185dbb
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:331fb2b7f8ecd022b2d839d7c7167894d42e426406c8fde67ed9a8d4c78176de
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91eae2fa3fc02ee15efef1884ec4e560ef910ecd5d82acb8671a8d4707b80a43
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:805137337f587aa7655dd5f599cea8ff7a9813cfb0f1bcf2d0b7dd019723b5ad
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db6a8aa79fcfebce03524e9b1b6b5ceec390e3bb9a2a198149bb646d26309c9b
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.5,
5
  "eval_steps": 500,
6
- "global_step": 7500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -532,6 +532,41 @@
532
  "learning_rate": 1.6668e-05,
533
  "loss": 0.0604,
534
  "step": 7500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
535
  }
536
  ],
537
  "logging_steps": 100,
@@ -551,7 +586,7 @@
551
  "attributes": {}
552
  }
553
  },
554
- "total_flos": 1.82687367168e+16,
555
  "train_batch_size": 4,
556
  "trial_name": null,
557
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.5333333333333333,
5
  "eval_steps": 500,
6
+ "global_step": 8000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
532
  "learning_rate": 1.6668e-05,
533
  "loss": 0.0604,
534
  "step": 7500
535
+ },
536
+ {
537
+ "epoch": 0.5066666666666667,
538
+ "grad_norm": 0.35880565643310547,
539
+ "learning_rate": 1.6623555555555556e-05,
540
+ "loss": 0.0685,
541
+ "step": 7600
542
+ },
543
+ {
544
+ "epoch": 0.5133333333333333,
545
+ "grad_norm": 0.12522092461585999,
546
+ "learning_rate": 1.6579111111111114e-05,
547
+ "loss": 0.0806,
548
+ "step": 7700
549
+ },
550
+ {
551
+ "epoch": 0.52,
552
+ "grad_norm": 0.1332361251115799,
553
+ "learning_rate": 1.653466666666667e-05,
554
+ "loss": 0.0699,
555
+ "step": 7800
556
+ },
557
+ {
558
+ "epoch": 0.5266666666666666,
559
+ "grad_norm": 0.2704123854637146,
560
+ "learning_rate": 1.6490222222222225e-05,
561
+ "loss": 0.0685,
562
+ "step": 7900
563
+ },
564
+ {
565
+ "epoch": 0.5333333333333333,
566
+ "grad_norm": 0.1266499161720276,
567
+ "learning_rate": 1.644577777777778e-05,
568
+ "loss": 0.0699,
569
+ "step": 8000
570
  }
571
  ],
572
  "logging_steps": 100,
 
586
  "attributes": {}
587
  }
588
  },
589
+ "total_flos": 1.948665249792e+16,
590
  "train_batch_size": 4,
591
  "trial_name": null,
592
  "trial_params": null