rakhman-llm commited on
Commit
f68c379
·
verified ·
1 Parent(s): e554cc1

Training in progress, step 33000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2ee30a202b5c4cf7fda53d0395699d121bcda28082c91bf553e9af5faf8a149b
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f3fa8a2349c8b7e8ef7591faf4142fae1ab2981c5397d9e3f22dbb692b3d52a
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e718a43f6e045c42f385b07d2ace5849f04af633001edd559d8966df49316ab3
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b42426cc36edf758e4d57f94cd488382d45ee0e94746d9273813cf1a72d6fbcc
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:230a3736d669f255b79c1310914eb307cf0cbcaffc29ea6c220c8c34f3f16376
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb9092d8d8eea8d4b54fe2ad3f4e9f22ac5e285340629fa61e0b094539adabee
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:628f3ee901d7a700b603c7233dc69a829b7b7b1f7aa2edec89c4540f453a5834
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5011c95b426af587e7306c32d3c809014be6f544ede16b0ee848cd35fd4378b6
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 2.1666666666666665,
5
  "eval_steps": 500,
6
- "global_step": 32500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -2298,6 +2298,41 @@
2298
  "learning_rate": 5.560888888888889e-06,
2299
  "loss": 0.0522,
2300
  "step": 32500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2301
  }
2302
  ],
2303
  "logging_steps": 100,
@@ -2317,7 +2352,7 @@
2317
  "attributes": {}
2318
  }
2319
  },
2320
- "total_flos": 7.91645257728e+16,
2321
  "train_batch_size": 4,
2322
  "trial_name": null,
2323
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 2.2,
5
  "eval_steps": 500,
6
+ "global_step": 33000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
2298
  "learning_rate": 5.560888888888889e-06,
2299
  "loss": 0.0522,
2300
  "step": 32500
2301
+ },
2302
+ {
2303
+ "epoch": 2.1733333333333333,
2304
+ "grad_norm": 0.21274752914905548,
2305
+ "learning_rate": 5.516444444444445e-06,
2306
+ "loss": 0.0523,
2307
+ "step": 32600
2308
+ },
2309
+ {
2310
+ "epoch": 2.18,
2311
+ "grad_norm": 0.3107999563217163,
2312
+ "learning_rate": 5.472e-06,
2313
+ "loss": 0.0526,
2314
+ "step": 32700
2315
+ },
2316
+ {
2317
+ "epoch": 2.1866666666666665,
2318
+ "grad_norm": 0.19043102860450745,
2319
+ "learning_rate": 5.4275555555555555e-06,
2320
+ "loss": 0.0522,
2321
+ "step": 32800
2322
+ },
2323
+ {
2324
+ "epoch": 2.1933333333333334,
2325
+ "grad_norm": 0.17905379831790924,
2326
+ "learning_rate": 5.383111111111112e-06,
2327
+ "loss": 0.052,
2328
+ "step": 32900
2329
+ },
2330
+ {
2331
+ "epoch": 2.2,
2332
+ "grad_norm": 0.3581301271915436,
2333
+ "learning_rate": 5.338666666666668e-06,
2334
+ "loss": 0.0522,
2335
+ "step": 33000
2336
  }
2337
  ],
2338
  "logging_steps": 100,
 
2352
  "attributes": {}
2353
  }
2354
  },
2355
+ "total_flos": 8.038244155392e+16,
2356
  "train_batch_size": 4,
2357
  "trial_name": null,
2358
  "trial_params": null