rakhman-llm commited on
Commit
38eee8d
·
verified ·
1 Parent(s): 7d301f1

Training in progress, step 7500, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3b0338daac15cfc77c983eea6a6122874f95852031cec6efb33fb6843974f909
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e017ff86b8a295f11193c500291f5991ea24bf880a065787121b3fb732878a01
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:eff7218517ca4458c3d29b7230cefb18dfe8c171f0a8309d5494771317c7b329
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8013718d7b486d4bea38838926eb12f405577f302b56db31ac78355ece57c039
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c8c94fe50942676a4b800b8807f5b832311548d8f2c40eff1078995f3e3f0ab5
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:331fb2b7f8ecd022b2d839d7c7167894d42e426406c8fde67ed9a8d4c78176de
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c8e5e69cb9645219d028edd3b028ebf30583f1263680ed0ab467551ed740d562
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:805137337f587aa7655dd5f599cea8ff7a9813cfb0f1bcf2d0b7dd019723b5ad
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.4666666666666667,
5
  "eval_steps": 500,
6
- "global_step": 7000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -497,6 +497,41 @@
497
  "learning_rate": 1.6890222222222224e-05,
498
  "loss": 0.067,
499
  "step": 7000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
500
  }
501
  ],
502
  "logging_steps": 100,
@@ -516,7 +551,7 @@
516
  "attributes": {}
517
  }
518
  },
519
- "total_flos": 1.705082093568e+16,
520
  "train_batch_size": 4,
521
  "trial_name": null,
522
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.5,
5
  "eval_steps": 500,
6
+ "global_step": 7500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
497
  "learning_rate": 1.6890222222222224e-05,
498
  "loss": 0.067,
499
  "step": 7000
500
+ },
501
+ {
502
+ "epoch": 0.47333333333333333,
503
+ "grad_norm": 0.21629302203655243,
504
+ "learning_rate": 1.684577777777778e-05,
505
+ "loss": 0.0684,
506
+ "step": 7100
507
+ },
508
+ {
509
+ "epoch": 0.48,
510
+ "grad_norm": 0.45061948895454407,
511
+ "learning_rate": 1.6801333333333335e-05,
512
+ "loss": 0.0713,
513
+ "step": 7200
514
+ },
515
+ {
516
+ "epoch": 0.4866666666666667,
517
+ "grad_norm": 0.19869284331798553,
518
+ "learning_rate": 1.675688888888889e-05,
519
+ "loss": 0.0608,
520
+ "step": 7300
521
+ },
522
+ {
523
+ "epoch": 0.49333333333333335,
524
+ "grad_norm": 0.23369118571281433,
525
+ "learning_rate": 1.6712444444444445e-05,
526
+ "loss": 0.0675,
527
+ "step": 7400
528
+ },
529
+ {
530
+ "epoch": 0.5,
531
+ "grad_norm": 0.16967739164829254,
532
+ "learning_rate": 1.6668e-05,
533
+ "loss": 0.0604,
534
+ "step": 7500
535
  }
536
  ],
537
  "logging_steps": 100,
 
551
  "attributes": {}
552
  }
553
  },
554
+ "total_flos": 1.82687367168e+16,
555
  "train_batch_size": 4,
556
  "trial_name": null,
557
  "trial_params": null