rakhman-llm commited on
Commit
9909aa6
·
verified ·
1 Parent(s): 557a2db

Training in progress, step 8500, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a43ee6984f5f990cd4d1d7d439666b7dfa7b56da1d97ae91a2c6727d6281cb2d
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d940b624e42af265966f792f158b5e6dff6cf498a4c59d70a1629ae2fdc2efc
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1f67eef48f1db3ec75565c0e239b37d2a1a24b92d26f253f611940b1cf185dbb
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a2bec32040a69e58023b5509b0fb90efe9794b8b9079e790fb6687d3ff55f17
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:91eae2fa3fc02ee15efef1884ec4e560ef910ecd5d82acb8671a8d4707b80a43
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08998c919fed10f9ef4173715a99abc07523cdf2197bb7756b598ec6ecad7017
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:db6a8aa79fcfebce03524e9b1b6b5ceec390e3bb9a2a198149bb646d26309c9b
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a2d7ef4474f599b4e657eaa32ef5df2560c025d5712ce31e544f317cb28ad99
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.5333333333333333,
5
  "eval_steps": 500,
6
- "global_step": 8000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -567,6 +567,41 @@
567
  "learning_rate": 1.644577777777778e-05,
568
  "loss": 0.0699,
569
  "step": 8000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
570
  }
571
  ],
572
  "logging_steps": 100,
@@ -586,7 +621,7 @@
586
  "attributes": {}
587
  }
588
  },
589
- "total_flos": 1.948665249792e+16,
590
  "train_batch_size": 4,
591
  "trial_name": null,
592
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.5666666666666667,
5
  "eval_steps": 500,
6
+ "global_step": 8500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
567
  "learning_rate": 1.644577777777778e-05,
568
  "loss": 0.0699,
569
  "step": 8000
570
+ },
571
+ {
572
+ "epoch": 0.54,
573
+ "grad_norm": 0.03966674581170082,
574
+ "learning_rate": 1.6401333333333335e-05,
575
+ "loss": 0.0634,
576
+ "step": 8100
577
+ },
578
+ {
579
+ "epoch": 0.5466666666666666,
580
+ "grad_norm": 0.25476112961769104,
581
+ "learning_rate": 1.635688888888889e-05,
582
+ "loss": 0.0725,
583
+ "step": 8200
584
+ },
585
+ {
586
+ "epoch": 0.5533333333333333,
587
+ "grad_norm": 0.25321516394615173,
588
+ "learning_rate": 1.6312444444444446e-05,
589
+ "loss": 0.0683,
590
+ "step": 8300
591
+ },
592
+ {
593
+ "epoch": 0.56,
594
+ "grad_norm": 0.25231412053108215,
595
+ "learning_rate": 1.6268e-05,
596
+ "loss": 0.0735,
597
+ "step": 8400
598
+ },
599
+ {
600
+ "epoch": 0.5666666666666667,
601
+ "grad_norm": 0.08408234268426895,
602
+ "learning_rate": 1.6223555555555556e-05,
603
+ "loss": 0.0677,
604
+ "step": 8500
605
  }
606
  ],
607
  "logging_steps": 100,
 
621
  "attributes": {}
622
  }
623
  },
624
+ "total_flos": 2.070456827904e+16,
625
  "train_batch_size": 4,
626
  "trial_name": null,
627
  "trial_params": null