rakhman-llm commited on
Commit
4724932
·
verified ·
1 Parent(s): ff0cff5

Training in progress, step 35000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:70be4ea2eb3b605b7bbf19f657f7030b3328a2472cfce6fd98277ee878f3d42d
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54c3b650aacadef454b8db2ad2f4748bd2cc9caa9988d9586a8661ac87876f39
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0ea42d9a3f547f1d5475d09c5cf077e380e3710c708660172e4c053a5a76213e
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3300d0ac109580f5456c1f0fcf9e3276c33da9efc5df83c8e4bc2c715849e9ac
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fd91d8604c3c73ee737187beb59c853ab84bc04334806ba928b483da59e07022
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c640612ca6874adba07e3028442a5de517f2bd82ce8dc7d5cdd1701c4e1b1ee5
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7b582570b847f10b59692f51f82170b4e892f7a8342116f22c17cd2ce4cb0c5e
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ec6550e2fb94624e4937a6e7d695ee415ac08cd206a873e9804651b6a923373
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 2.3,
5
  "eval_steps": 500,
6
- "global_step": 34500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -2438,6 +2438,41 @@
2438
  "learning_rate": 4.672444444444445e-06,
2439
  "loss": 0.0484,
2440
  "step": 34500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2441
  }
2442
  ],
2443
  "logging_steps": 100,
@@ -2457,7 +2492,7 @@
2457
  "attributes": {}
2458
  }
2459
  },
2460
- "total_flos": 8.403618889728e+16,
2461
  "train_batch_size": 4,
2462
  "trial_name": null,
2463
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 2.3333333333333335,
5
  "eval_steps": 500,
6
+ "global_step": 35000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
2438
  "learning_rate": 4.672444444444445e-06,
2439
  "loss": 0.0484,
2440
  "step": 34500
2441
+ },
2442
+ {
2443
+ "epoch": 2.3066666666666666,
2444
+ "grad_norm": 0.112776979804039,
2445
+ "learning_rate": 4.628e-06,
2446
+ "loss": 0.0534,
2447
+ "step": 34600
2448
+ },
2449
+ {
2450
+ "epoch": 2.3133333333333335,
2451
+ "grad_norm": 0.3088347315788269,
2452
+ "learning_rate": 4.5835555555555565e-06,
2453
+ "loss": 0.0565,
2454
+ "step": 34700
2455
+ },
2456
+ {
2457
+ "epoch": 2.32,
2458
+ "grad_norm": 0.18766255676746368,
2459
+ "learning_rate": 4.539111111111112e-06,
2460
+ "loss": 0.0548,
2461
+ "step": 34800
2462
+ },
2463
+ {
2464
+ "epoch": 2.3266666666666667,
2465
+ "grad_norm": 0.24050170183181763,
2466
+ "learning_rate": 4.494666666666667e-06,
2467
+ "loss": 0.0602,
2468
+ "step": 34900
2469
+ },
2470
+ {
2471
+ "epoch": 2.3333333333333335,
2472
+ "grad_norm": 0.15460790693759918,
2473
+ "learning_rate": 4.450222222222222e-06,
2474
+ "loss": 0.0558,
2475
+ "step": 35000
2476
  }
2477
  ],
2478
  "logging_steps": 100,
 
2492
  "attributes": {}
2493
  }
2494
  },
2495
+ "total_flos": 8.52541046784e+16,
2496
  "train_batch_size": 4,
2497
  "trial_name": null,
2498
  "trial_params": null