rakhman-llm commited on
Commit
5dc894e
·
verified ·
1 Parent(s): 5422dba

Training in progress, step 29000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3614617ee76c86000db29cd744353eb118d9a3d06e5a97f4af361b3577a0b872
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d0d7856ac14e3c9a690e3c9dee53a7a64870218a71216a1c30e598f861133e14
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e5ba13f67cae6dd4ea0abf5c768c7bd8aaa350d6c17a47580822772f8d79c646
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed3cf9b04c05bb73a17b1d26a6bf42f7dfd71e2f3a8b717e1cbc8e86af15e592
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:606d09e46c381a6ca79c1c2761e1f43da7a02f07449ea8dc08c4fd55302f6057
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13275cc6296652d5dbe96fab3538f55b506c905c768d5c0e2ff0cc00f18629a9
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cf940af0b3fe4a79522e2cb3cee03ca9bdc7bb9e1119203316379171ea473573
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6cce86ca086e9ceba19d07f438b59e29c1c4e8071fe80d69909be3a44b7f8d02
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 1.9,
5
  "eval_steps": 500,
6
- "global_step": 28500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -2010,6 +2010,41 @@
2010
  "learning_rate": 7.337777777777778e-06,
2011
  "loss": 0.0541,
2012
  "step": 28500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2013
  }
2014
  ],
2015
  "logging_steps": 100,
@@ -2029,7 +2064,7 @@
2029
  "attributes": {}
2030
  }
2031
  },
2032
- "total_flos": 6.942119952384e+16,
2033
  "train_batch_size": 4,
2034
  "trial_name": null,
2035
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 1.9333333333333333,
5
  "eval_steps": 500,
6
+ "global_step": 29000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
2010
  "learning_rate": 7.337777777777778e-06,
2011
  "loss": 0.0541,
2012
  "step": 28500
2013
+ },
2014
+ {
2015
+ "epoch": 1.9066666666666667,
2016
+ "grad_norm": 0.15182781219482422,
2017
+ "learning_rate": 7.2933333333333335e-06,
2018
+ "loss": 0.0566,
2019
+ "step": 28600
2020
+ },
2021
+ {
2022
+ "epoch": 1.9133333333333333,
2023
+ "grad_norm": 0.18824052810668945,
2024
+ "learning_rate": 7.24888888888889e-06,
2025
+ "loss": 0.0566,
2026
+ "step": 28700
2027
+ },
2028
+ {
2029
+ "epoch": 1.92,
2030
+ "grad_norm": 0.2716074585914612,
2031
+ "learning_rate": 7.204444444444445e-06,
2032
+ "loss": 0.0571,
2033
+ "step": 28800
2034
+ },
2035
+ {
2036
+ "epoch": 1.9266666666666667,
2037
+ "grad_norm": 0.18709616363048553,
2038
+ "learning_rate": 7.16e-06,
2039
+ "loss": 0.0569,
2040
+ "step": 28900
2041
+ },
2042
+ {
2043
+ "epoch": 1.9333333333333333,
2044
+ "grad_norm": 0.1409674882888794,
2045
+ "learning_rate": 7.115555555555557e-06,
2046
+ "loss": 0.0603,
2047
+ "step": 29000
2048
  }
2049
  ],
2050
  "logging_steps": 100,
 
2064
  "attributes": {}
2065
  }
2066
  },
2067
+ "total_flos": 7.063911530496e+16,
2068
  "train_batch_size": 4,
2069
  "trial_name": null,
2070
  "trial_params": null