rakhman-llm commited on
Commit
ae8a8f1
·
verified ·
1 Parent(s): 47f4ff0

Training in progress, step 28500, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ff193c75cfe266a4454f3b865f678cd0068e31d158d56f3a08c31bd1c8ca8180
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3614617ee76c86000db29cd744353eb118d9a3d06e5a97f4af361b3577a0b872
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:058172c34ee10347ff67ff2341a5f4bfdcee6c70e269830728adf925c422cbcb
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e5ba13f67cae6dd4ea0abf5c768c7bd8aaa350d6c17a47580822772f8d79c646
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0242710dd76fcfc259b7718df92653875f59a11b4f0133d9bd37c3e685163566
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:606d09e46c381a6ca79c1c2761e1f43da7a02f07449ea8dc08c4fd55302f6057
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:117137eaff655ce606b81582da9e0c028be111ad40ac1072d2fad13931816255
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf940af0b3fe4a79522e2cb3cee03ca9bdc7bb9e1119203316379171ea473573
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 1.8666666666666667,
5
  "eval_steps": 500,
6
- "global_step": 28000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1975,6 +1975,41 @@
1975
  "learning_rate": 7.5595555555555565e-06,
1976
  "loss": 0.056,
1977
  "step": 28000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1978
  }
1979
  ],
1980
  "logging_steps": 100,
@@ -1994,7 +2029,7 @@
1994
  "attributes": {}
1995
  }
1996
  },
1997
- "total_flos": 6.820328374272e+16,
1998
  "train_batch_size": 4,
1999
  "trial_name": null,
2000
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 1.9,
5
  "eval_steps": 500,
6
+ "global_step": 28500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1975
  "learning_rate": 7.5595555555555565e-06,
1976
  "loss": 0.056,
1977
  "step": 28000
1978
+ },
1979
+ {
1980
+ "epoch": 1.8733333333333333,
1981
+ "grad_norm": 0.38631588220596313,
1982
+ "learning_rate": 7.515111111111112e-06,
1983
+ "loss": 0.0554,
1984
+ "step": 28100
1985
+ },
1986
+ {
1987
+ "epoch": 1.88,
1988
+ "grad_norm": 0.3035646677017212,
1989
+ "learning_rate": 7.470666666666667e-06,
1990
+ "loss": 0.0554,
1991
+ "step": 28200
1992
+ },
1993
+ {
1994
+ "epoch": 1.8866666666666667,
1995
+ "grad_norm": 0.2283605933189392,
1996
+ "learning_rate": 7.426222222222222e-06,
1997
+ "loss": 0.0584,
1998
+ "step": 28300
1999
+ },
2000
+ {
2001
+ "epoch": 1.8933333333333333,
2002
+ "grad_norm": 0.10160692036151886,
2003
+ "learning_rate": 7.381777777777779e-06,
2004
+ "loss": 0.0538,
2005
+ "step": 28400
2006
+ },
2007
+ {
2008
+ "epoch": 1.9,
2009
+ "grad_norm": 0.704264223575592,
2010
+ "learning_rate": 7.337777777777778e-06,
2011
+ "loss": 0.0541,
2012
+ "step": 28500
2013
  }
2014
  ],
2015
  "logging_steps": 100,
 
2029
  "attributes": {}
2030
  }
2031
  },
2032
+ "total_flos": 6.942119952384e+16,
2033
  "train_batch_size": 4,
2034
  "trial_name": null,
2035
  "trial_params": null