rakhman-llm commited on
Commit
64b48be
·
verified ·
1 Parent(s): 42f05a9

Training in progress, step 5000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8fdbb9af804d15f5110973d90186ec84defa2a665a8482fd83c11a269501602d
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ae5c81c1c60500c4a5e1dd6dc813ec1e98722c56c518f3879627e98375fc97c
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6e8eed0cff93a66dd9f7dc35bf13769c84be7bd70a4333fe001f86cc9ee212ba
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02ec7dd85fee160cc36f0c35de11d3c1f3f62be09dc2acdbea1436d93df16a1f
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2ca0d52bc1b71ba08e7156c2b45d758a196d38988c6eda67b4df0e9d832fe78c
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e395ce6518143962b06454b53148180fd527e3c70ad1a04b860142fa90882881
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:574db7da084c9454e4a06ea3487f7cdcb4dd46faa4038c9a6b4a085fb43bd4b6
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:133e76e86cd14a96bc531992fc191ed101dcf5373cd51d0b24c7b1c479a2d7b7
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.3,
5
  "eval_steps": 500,
6
- "global_step": 4500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -322,6 +322,41 @@
322
  "learning_rate": 1.800088888888889e-05,
323
  "loss": 0.0752,
324
  "step": 4500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
325
  }
326
  ],
327
  "logging_steps": 100,
@@ -341,7 +376,7 @@
341
  "attributes": {}
342
  }
343
  },
344
- "total_flos": 1.096124203008e+16,
345
  "train_batch_size": 4,
346
  "trial_name": null,
347
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.3333333333333333,
5
  "eval_steps": 500,
6
+ "global_step": 5000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
322
  "learning_rate": 1.800088888888889e-05,
323
  "loss": 0.0752,
324
  "step": 4500
325
+ },
326
+ {
327
+ "epoch": 0.30666666666666664,
328
+ "grad_norm": 0.22465260326862335,
329
+ "learning_rate": 1.7956444444444445e-05,
330
+ "loss": 0.068,
331
+ "step": 4600
332
+ },
333
+ {
334
+ "epoch": 0.31333333333333335,
335
+ "grad_norm": 0.0787540078163147,
336
+ "learning_rate": 1.7912e-05,
337
+ "loss": 0.077,
338
+ "step": 4700
339
+ },
340
+ {
341
+ "epoch": 0.32,
342
+ "grad_norm": 0.2305675595998764,
343
+ "learning_rate": 1.786755555555556e-05,
344
+ "loss": 0.0727,
345
+ "step": 4800
346
+ },
347
+ {
348
+ "epoch": 0.32666666666666666,
349
+ "grad_norm": 0.5901878476142883,
350
+ "learning_rate": 1.7823111111111114e-05,
351
+ "loss": 0.0717,
352
+ "step": 4900
353
+ },
354
+ {
355
+ "epoch": 0.3333333333333333,
356
+ "grad_norm": 0.25235211849212646,
357
+ "learning_rate": 1.777866666666667e-05,
358
+ "loss": 0.0669,
359
+ "step": 5000
360
  }
361
  ],
362
  "logging_steps": 100,
 
376
  "attributes": {}
377
  }
378
  },
379
+ "total_flos": 1.21791578112e+16,
380
  "train_batch_size": 4,
381
  "trial_name": null,
382
  "trial_params": null