rakhman-llm commited on
Commit
bbde5f2
·
verified ·
1 Parent(s): b143bed

Training in progress, step 35500, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:54c3b650aacadef454b8db2ad2f4748bd2cc9caa9988d9586a8661ac87876f39
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37250fd25ac520b5292199088ef27546da88796721ea4eb40695113400ccfa7b
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3300d0ac109580f5456c1f0fcf9e3276c33da9efc5df83c8e4bc2c715849e9ac
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad6751a841c8864f7c7c1ceb77b425b700e73e59e70f845d753c55a66facc1b6
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c640612ca6874adba07e3028442a5de517f2bd82ce8dc7d5cdd1701c4e1b1ee5
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2920a601864c0632a370a1ac1dcdad5f1e27e566971eb45ef713c65fc1086f00
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1ec6550e2fb94624e4937a6e7d695ee415ac08cd206a873e9804651b6a923373
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0cc3d0c5d65fdec57f5db591d38345123ff409f36daf00beadf6d67548eb0bd8
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 2.3333333333333335,
5
  "eval_steps": 500,
6
- "global_step": 35000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -2473,6 +2473,41 @@
2473
  "learning_rate": 4.450222222222222e-06,
2474
  "loss": 0.0558,
2475
  "step": 35000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2476
  }
2477
  ],
2478
  "logging_steps": 100,
@@ -2492,7 +2527,7 @@
2492
  "attributes": {}
2493
  }
2494
  },
2495
- "total_flos": 8.52541046784e+16,
2496
  "train_batch_size": 4,
2497
  "trial_name": null,
2498
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 2.3666666666666667,
5
  "eval_steps": 500,
6
+ "global_step": 35500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
2473
  "learning_rate": 4.450222222222222e-06,
2474
  "loss": 0.0558,
2475
  "step": 35000
2476
+ },
2477
+ {
2478
+ "epoch": 2.34,
2479
+ "grad_norm": 0.23494431376457214,
2480
+ "learning_rate": 4.405777777777778e-06,
2481
+ "loss": 0.0521,
2482
+ "step": 35100
2483
+ },
2484
+ {
2485
+ "epoch": 2.3466666666666667,
2486
+ "grad_norm": 0.19245880842208862,
2487
+ "learning_rate": 4.3613333333333335e-06,
2488
+ "loss": 0.0516,
2489
+ "step": 35200
2490
+ },
2491
+ {
2492
+ "epoch": 2.3533333333333335,
2493
+ "grad_norm": 0.30625805258750916,
2494
+ "learning_rate": 4.3173333333333336e-06,
2495
+ "loss": 0.0519,
2496
+ "step": 35300
2497
+ },
2498
+ {
2499
+ "epoch": 2.36,
2500
+ "grad_norm": 0.30782145261764526,
2501
+ "learning_rate": 4.272888888888889e-06,
2502
+ "loss": 0.0495,
2503
+ "step": 35400
2504
+ },
2505
+ {
2506
+ "epoch": 2.3666666666666667,
2507
+ "grad_norm": 0.25636398792266846,
2508
+ "learning_rate": 4.228444444444445e-06,
2509
+ "loss": 0.0529,
2510
+ "step": 35500
2511
  }
2512
  ],
2513
  "logging_steps": 100,
 
2527
  "attributes": {}
2528
  }
2529
  },
2530
+ "total_flos": 8.647202045952e+16,
2531
  "train_batch_size": 4,
2532
  "trial_name": null,
2533
  "trial_params": null