rakhman-llm commited on
Commit
2a249d8
·
verified ·
1 Parent(s): b4fd620

Training in progress, step 36500, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:61dce64509302181dbd0102d94f8e63119a5e468f1ab30981e6b95a40a16a0f7
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66629fbba7f2d42de83407ccc7c4a7cb20d143079ca2ed45dbc07d04a90a40bf
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8aee7a7bb130bc58eccba7310fd2e1c0ff9bc9a9460551dab2ac36a3c09f233f
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d02b75121788cccbde8e0ee668a0edad332b5ebb58494b81e48d2fa78f15f7d
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d52b5d54b1b44c26afdd58b8aa610c5f355b531f51747715e0d2588d67b0dd03
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e9822f3fcdc1597872af5dd87f19db7e35ac92105cd8babdb7bf1d6ea8c5abd
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3545c32cd08bfe14c5940277cc90a883188128e319d14e37ac1ae89581718ce4
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e696025a153a4c23b525b445028cbabae1500984cfdb159ce8be1d5922b1b1ea
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 2.4,
5
  "eval_steps": 500,
6
- "global_step": 36000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -2543,6 +2543,41 @@
2543
  "learning_rate": 4.006222222222223e-06,
2544
  "loss": 0.0523,
2545
  "step": 36000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2546
  }
2547
  ],
2548
  "logging_steps": 100,
@@ -2562,7 +2597,7 @@
2562
  "attributes": {}
2563
  }
2564
  },
2565
- "total_flos": 8.768993624064e+16,
2566
  "train_batch_size": 4,
2567
  "trial_name": null,
2568
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 2.4333333333333336,
5
  "eval_steps": 500,
6
+ "global_step": 36500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
2543
  "learning_rate": 4.006222222222223e-06,
2544
  "loss": 0.0523,
2545
  "step": 36000
2546
+ },
2547
+ {
2548
+ "epoch": 2.4066666666666667,
2549
+ "grad_norm": 0.1548595130443573,
2550
+ "learning_rate": 3.961777777777778e-06,
2551
+ "loss": 0.0498,
2552
+ "step": 36100
2553
+ },
2554
+ {
2555
+ "epoch": 2.413333333333333,
2556
+ "grad_norm": 0.13865716755390167,
2557
+ "learning_rate": 3.917333333333333e-06,
2558
+ "loss": 0.0522,
2559
+ "step": 36200
2560
+ },
2561
+ {
2562
+ "epoch": 2.42,
2563
+ "grad_norm": 0.20406047999858856,
2564
+ "learning_rate": 3.872888888888889e-06,
2565
+ "loss": 0.0481,
2566
+ "step": 36300
2567
+ },
2568
+ {
2569
+ "epoch": 2.4266666666666667,
2570
+ "grad_norm": 0.20818284153938293,
2571
+ "learning_rate": 3.8284444444444445e-06,
2572
+ "loss": 0.0524,
2573
+ "step": 36400
2574
+ },
2575
+ {
2576
+ "epoch": 2.4333333333333336,
2577
+ "grad_norm": 0.37012940645217896,
2578
+ "learning_rate": 3.7840000000000005e-06,
2579
+ "loss": 0.0445,
2580
+ "step": 36500
2581
  }
2582
  ],
2583
  "logging_steps": 100,
 
2597
  "attributes": {}
2598
  }
2599
  },
2600
+ "total_flos": 8.890785202176e+16,
2601
  "train_batch_size": 4,
2602
  "trial_name": null,
2603
  "trial_params": null