rakhman-llm commited on
Commit
65d2cc1
·
verified ·
1 Parent(s): b568abc

Training in progress, step 33500, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4f3fa8a2349c8b7e8ef7591faf4142fae1ab2981c5397d9e3f22dbb692b3d52a
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b97182c688d993733fe30b3cd254e6701a753aa3cc2386b86d587fdb03bb6673
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b42426cc36edf758e4d57f94cd488382d45ee0e94746d9273813cf1a72d6fbcc
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7856bcf04b20cd8307ae2c76430eb38324a8c0698517e73495150d5bd27dff1b
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cb9092d8d8eea8d4b54fe2ad3f4e9f22ac5e285340629fa61e0b094539adabee
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:137247deb099a0895b19ce9e4c6b96a673b81c10a6ea5cdab1f3338408e77892
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5011c95b426af587e7306c32d3c809014be6f544ede16b0ee848cd35fd4378b6
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e4fd446a83e934c2d5675c18ca5f4bc5dada518d6ba6f33540a2d77e70605ffe
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 2.2,
5
  "eval_steps": 500,
6
- "global_step": 33000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -2333,6 +2333,41 @@
2333
  "learning_rate": 5.338666666666668e-06,
2334
  "loss": 0.0522,
2335
  "step": 33000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2336
  }
2337
  ],
2338
  "logging_steps": 100,
@@ -2352,7 +2387,7 @@
2352
  "attributes": {}
2353
  }
2354
  },
2355
- "total_flos": 8.038244155392e+16,
2356
  "train_batch_size": 4,
2357
  "trial_name": null,
2358
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 2.2333333333333334,
5
  "eval_steps": 500,
6
+ "global_step": 33500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
2333
  "learning_rate": 5.338666666666668e-06,
2334
  "loss": 0.0522,
2335
  "step": 33000
2336
+ },
2337
+ {
2338
+ "epoch": 2.2066666666666666,
2339
+ "grad_norm": 0.1424885094165802,
2340
+ "learning_rate": 5.294222222222223e-06,
2341
+ "loss": 0.0542,
2342
+ "step": 33100
2343
+ },
2344
+ {
2345
+ "epoch": 2.2133333333333334,
2346
+ "grad_norm": 0.29376089572906494,
2347
+ "learning_rate": 5.249777777777778e-06,
2348
+ "loss": 0.0466,
2349
+ "step": 33200
2350
+ },
2351
+ {
2352
+ "epoch": 2.22,
2353
+ "grad_norm": 0.21891391277313232,
2354
+ "learning_rate": 5.205333333333333e-06,
2355
+ "loss": 0.051,
2356
+ "step": 33300
2357
+ },
2358
+ {
2359
+ "epoch": 2.2266666666666666,
2360
+ "grad_norm": 0.09811867028474808,
2361
+ "learning_rate": 5.1608888888888894e-06,
2362
+ "loss": 0.053,
2363
+ "step": 33400
2364
+ },
2365
+ {
2366
+ "epoch": 2.2333333333333334,
2367
+ "grad_norm": 0.17383785545825958,
2368
+ "learning_rate": 5.116444444444445e-06,
2369
+ "loss": 0.0486,
2370
+ "step": 33500
2371
  }
2372
  ],
2373
  "logging_steps": 100,
 
2387
  "attributes": {}
2388
  }
2389
  },
2390
+ "total_flos": 8.160035733504e+16,
2391
  "train_batch_size": 4,
2392
  "trial_name": null,
2393
  "trial_params": null