rakhman-llm commited on
Commit
1d1c867
·
verified ·
1 Parent(s): 89b309c

Training in progress, step 34000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b97182c688d993733fe30b3cd254e6701a753aa3cc2386b86d587fdb03bb6673
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e4ab360c914af5780dbdab371c96384f68fadc50929f69d6d527fb768d4c4fc
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7856bcf04b20cd8307ae2c76430eb38324a8c0698517e73495150d5bd27dff1b
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:efe48f4318badae6c7301c7e1af6b474ccbe7f71331abdb7cdd8a9a8ea72715a
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:137247deb099a0895b19ce9e4c6b96a673b81c10a6ea5cdab1f3338408e77892
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6d089c8bc4767a68a0971e6301211196ccb288fdbf802f4d0800d2c395bd6e0
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e4fd446a83e934c2d5675c18ca5f4bc5dada518d6ba6f33540a2d77e70605ffe
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5881a400174694043dc48b7e6a0f928fe1bcddfe30dc261df5a9e4e69d0e718
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 2.2333333333333334,
5
  "eval_steps": 500,
6
- "global_step": 33500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -2368,6 +2368,41 @@
2368
  "learning_rate": 5.116444444444445e-06,
2369
  "loss": 0.0486,
2370
  "step": 33500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2371
  }
2372
  ],
2373
  "logging_steps": 100,
@@ -2387,7 +2422,7 @@
2387
  "attributes": {}
2388
  }
2389
  },
2390
- "total_flos": 8.160035733504e+16,
2391
  "train_batch_size": 4,
2392
  "trial_name": null,
2393
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 2.2666666666666666,
5
  "eval_steps": 500,
6
+ "global_step": 34000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
2368
  "learning_rate": 5.116444444444445e-06,
2369
  "loss": 0.0486,
2370
  "step": 33500
2371
+ },
2372
+ {
2373
+ "epoch": 2.24,
2374
+ "grad_norm": 0.20023219287395477,
2375
+ "learning_rate": 5.072e-06,
2376
+ "loss": 0.0509,
2377
+ "step": 33600
2378
+ },
2379
+ {
2380
+ "epoch": 2.2466666666666666,
2381
+ "grad_norm": 0.24713067710399628,
2382
+ "learning_rate": 5.027555555555555e-06,
2383
+ "loss": 0.0487,
2384
+ "step": 33700
2385
+ },
2386
+ {
2387
+ "epoch": 2.2533333333333334,
2388
+ "grad_norm": 0.2585085928440094,
2389
+ "learning_rate": 4.983111111111111e-06,
2390
+ "loss": 0.0464,
2391
+ "step": 33800
2392
+ },
2393
+ {
2394
+ "epoch": 2.26,
2395
+ "grad_norm": 0.10687225311994553,
2396
+ "learning_rate": 4.938666666666667e-06,
2397
+ "loss": 0.0558,
2398
+ "step": 33900
2399
+ },
2400
+ {
2401
+ "epoch": 2.2666666666666666,
2402
+ "grad_norm": 0.2748579680919647,
2403
+ "learning_rate": 4.8942222222222225e-06,
2404
+ "loss": 0.0533,
2405
+ "step": 34000
2406
  }
2407
  ],
2408
  "logging_steps": 100,
 
2422
  "attributes": {}
2423
  }
2424
  },
2425
+ "total_flos": 8.281827311616e+16,
2426
  "train_batch_size": 4,
2427
  "trial_name": null,
2428
  "trial_params": null