rakhman-llm commited on
Commit
7b67ee9
·
verified ·
1 Parent(s): f747f68

Training in progress, step 37000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:66629fbba7f2d42de83407ccc7c4a7cb20d143079ca2ed45dbc07d04a90a40bf
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0fd691e92552d902e006c1dbd51666fbe879b2be04760327e9ca887670be688f
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6d02b75121788cccbde8e0ee668a0edad332b5ebb58494b81e48d2fa78f15f7d
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f14ad18c691195cd15268660ba702ce3fc7337f8480e501b361bd9677af8b07e
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6e9822f3fcdc1597872af5dd87f19db7e35ac92105cd8babdb7bf1d6ea8c5abd
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:810ed5e7c2ae1b96c2bbc38902777e6779874b7cdcf097af17b2123a3bf9831a
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e696025a153a4c23b525b445028cbabae1500984cfdb159ce8be1d5922b1b1ea
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f866da3a9c3c847bb3514492ca12f1bdef65b499f47966b6692e65b75559e3e5
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 2.4333333333333336,
5
  "eval_steps": 500,
6
- "global_step": 36500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -2578,6 +2578,41 @@
2578
  "learning_rate": 3.7840000000000005e-06,
2579
  "loss": 0.0445,
2580
  "step": 36500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2581
  }
2582
  ],
2583
  "logging_steps": 100,
@@ -2597,7 +2632,7 @@
2597
  "attributes": {}
2598
  }
2599
  },
2600
- "total_flos": 8.890785202176e+16,
2601
  "train_batch_size": 4,
2602
  "trial_name": null,
2603
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 2.466666666666667,
5
  "eval_steps": 500,
6
+ "global_step": 37000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
2578
  "learning_rate": 3.7840000000000005e-06,
2579
  "loss": 0.0445,
2580
  "step": 36500
2581
+ },
2582
+ {
2583
+ "epoch": 2.44,
2584
+ "grad_norm": 0.07145357877016068,
2585
+ "learning_rate": 3.7395555555555558e-06,
2586
+ "loss": 0.0478,
2587
+ "step": 36600
2588
+ },
2589
+ {
2590
+ "epoch": 2.4466666666666668,
2591
+ "grad_norm": 0.12996427714824677,
2592
+ "learning_rate": 3.6951111111111114e-06,
2593
+ "loss": 0.0501,
2594
+ "step": 36700
2595
+ },
2596
+ {
2597
+ "epoch": 2.453333333333333,
2598
+ "grad_norm": 0.13270486891269684,
2599
+ "learning_rate": 3.6506666666666666e-06,
2600
+ "loss": 0.0542,
2601
+ "step": 36800
2602
+ },
2603
+ {
2604
+ "epoch": 2.46,
2605
+ "grad_norm": 0.18757346272468567,
2606
+ "learning_rate": 3.6062222222222227e-06,
2607
+ "loss": 0.0522,
2608
+ "step": 36900
2609
+ },
2610
+ {
2611
+ "epoch": 2.466666666666667,
2612
+ "grad_norm": 0.2180936336517334,
2613
+ "learning_rate": 3.561777777777778e-06,
2614
+ "loss": 0.0522,
2615
+ "step": 37000
2616
  }
2617
  ],
2618
  "logging_steps": 100,
 
2632
  "attributes": {}
2633
  }
2634
  },
2635
+ "total_flos": 9.012576780288e+16,
2636
  "train_batch_size": 4,
2637
  "trial_name": null,
2638
  "trial_params": null