rakhman-llm commited on
Commit
cca8b13
·
verified ·
1 Parent(s): c2e3d71

Training in progress, step 32500, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d9a244e9cb55adf7cda56ccb82e1a9bee92d342f231471fb5502ab29003b5e6b
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ee30a202b5c4cf7fda53d0395699d121bcda28082c91bf553e9af5faf8a149b
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f2fb02fb3109cf50329151281ddbb10cbbd67d7bcacbec12c415183618e7c8c4
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e718a43f6e045c42f385b07d2ace5849f04af633001edd559d8966df49316ab3
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ba3afeb76c9729f93ec4d0e7011fb72d06e1cea76b77b26bfd4a5e76df22f8c2
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:230a3736d669f255b79c1310914eb307cf0cbcaffc29ea6c220c8c34f3f16376
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:89d99968826ce47bd88d7b05d470f4fb17f0bcc9ac13362244060de99696b592
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:628f3ee901d7a700b603c7233dc69a829b7b7b1f7aa2edec89c4540f453a5834
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 2.1333333333333333,
5
  "eval_steps": 500,
6
- "global_step": 32000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -2263,6 +2263,41 @@
2263
  "learning_rate": 5.782666666666667e-06,
2264
  "loss": 0.0466,
2265
  "step": 32000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2266
  }
2267
  ],
2268
  "logging_steps": 100,
@@ -2282,7 +2317,7 @@
2282
  "attributes": {}
2283
  }
2284
  },
2285
- "total_flos": 7.794660999168e+16,
2286
  "train_batch_size": 4,
2287
  "trial_name": null,
2288
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 2.1666666666666665,
5
  "eval_steps": 500,
6
+ "global_step": 32500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
2263
  "learning_rate": 5.782666666666667e-06,
2264
  "loss": 0.0466,
2265
  "step": 32000
2266
+ },
2267
+ {
2268
+ "epoch": 2.14,
2269
+ "grad_norm": 0.2384709119796753,
2270
+ "learning_rate": 5.738222222222223e-06,
2271
+ "loss": 0.0567,
2272
+ "step": 32100
2273
+ },
2274
+ {
2275
+ "epoch": 2.1466666666666665,
2276
+ "grad_norm": 0.21881048381328583,
2277
+ "learning_rate": 5.6937777777777785e-06,
2278
+ "loss": 0.0577,
2279
+ "step": 32200
2280
+ },
2281
+ {
2282
+ "epoch": 2.1533333333333333,
2283
+ "grad_norm": 0.1541258692741394,
2284
+ "learning_rate": 5.649333333333334e-06,
2285
+ "loss": 0.0508,
2286
+ "step": 32300
2287
+ },
2288
+ {
2289
+ "epoch": 2.16,
2290
+ "grad_norm": 0.23964588344097137,
2291
+ "learning_rate": 5.605333333333334e-06,
2292
+ "loss": 0.0627,
2293
+ "step": 32400
2294
+ },
2295
+ {
2296
+ "epoch": 2.1666666666666665,
2297
+ "grad_norm": 0.27338939905166626,
2298
+ "learning_rate": 5.560888888888889e-06,
2299
+ "loss": 0.0522,
2300
+ "step": 32500
2301
  }
2302
  ],
2303
  "logging_steps": 100,
 
2317
  "attributes": {}
2318
  }
2319
  },
2320
+ "total_flos": 7.91645257728e+16,
2321
  "train_batch_size": 4,
2322
  "trial_name": null,
2323
  "trial_params": null