rakhman-llm commited on
Commit
b2d27cd
·
verified ·
1 Parent(s): c3a2595

Training in progress, step 4000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8153813283c4389ef5a9863dfbf03bd90de9af06723c2ed7469560d2f6cb9016
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75a0187432d05655fa95e9e4af04e889f2b84a4214d4995b5dcbc3829f6fa995
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3e3c6d028cc61eb16c10cfeb96ae472b416bf61c2430d616409a7e075447463c
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7eb7dc038fae5860bd92fb49fcb23e213f7099623915ac6963ae71d0e53c7159
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b21a8df05d6c8e0e6fd608ca76fd60fcfc2fde098551b903447e393817425942
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e2f6e8ef6ff0cccff7f602f9c9f7831d1edbb8062fe6f8d796324be0b7257b7f
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d3da069d23e3fc76f1f7013dd495a2b0f4544633e580b179d33500b4d99b2575
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6be5c125c7805acd88ceb3e79bd43e0a5bfe5ffe9cebeb5d64fd8beda77909cd
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.23333333333333334,
5
  "eval_steps": 500,
6
- "global_step": 3500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -252,6 +252,41 @@
252
  "learning_rate": 1.8445333333333334e-05,
253
  "loss": 0.0726,
254
  "step": 3500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
255
  }
256
  ],
257
  "logging_steps": 100,
@@ -271,7 +306,7 @@
271
  "attributes": {}
272
  }
273
  },
274
- "total_flos": 8525410467840000.0,
275
  "train_batch_size": 4,
276
  "trial_name": null,
277
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.26666666666666666,
5
  "eval_steps": 500,
6
+ "global_step": 4000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
252
  "learning_rate": 1.8445333333333334e-05,
253
  "loss": 0.0726,
254
  "step": 3500
255
+ },
256
+ {
257
+ "epoch": 0.24,
258
+ "grad_norm": 0.27142712473869324,
259
+ "learning_rate": 1.840088888888889e-05,
260
+ "loss": 0.0763,
261
+ "step": 3600
262
+ },
263
+ {
264
+ "epoch": 0.24666666666666667,
265
+ "grad_norm": 0.3764208257198334,
266
+ "learning_rate": 1.8356444444444445e-05,
267
+ "loss": 0.0816,
268
+ "step": 3700
269
+ },
270
+ {
271
+ "epoch": 0.25333333333333335,
272
+ "grad_norm": 0.2864600718021393,
273
+ "learning_rate": 1.8312e-05,
274
+ "loss": 0.0707,
275
+ "step": 3800
276
+ },
277
+ {
278
+ "epoch": 0.26,
279
+ "grad_norm": 0.2760820686817169,
280
+ "learning_rate": 1.8267555555555555e-05,
281
+ "loss": 0.0759,
282
+ "step": 3900
283
+ },
284
+ {
285
+ "epoch": 0.26666666666666666,
286
+ "grad_norm": 0.14416909217834473,
287
+ "learning_rate": 1.8223111111111114e-05,
288
+ "loss": 0.0621,
289
+ "step": 4000
290
  }
291
  ],
292
  "logging_steps": 100,
 
306
  "attributes": {}
307
  }
308
  },
309
+ "total_flos": 9743326248960000.0,
310
  "train_batch_size": 4,
311
  "trial_name": null,
312
  "trial_params": null