rakhman-llm commited on
Commit
209f2bd
·
verified ·
1 Parent(s): d36d59c

Training in progress, step 43500, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c0ecd85397c9873033c8fc72b8b958ac68dabe4479b23d12cd143f2bdaec2e49
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3f3f60f837bc2d6b892b2d0546640ca46d3f194d47c20345c8767c8ccaa5c07
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ba0de08f17477673f2db27ad78dc1cb0fb353da11bb179b4a06c1a9b0f4eb722
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6211545bbfeb1c43922e986aa4087c1ebccc0eedc2e220fb0dba4944f73f39c0
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9125611692f04b4e4f7334517ac150a7a29269d10071636ff9110a9581f5f331
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85869ac30f3cee5a03da0dea6f3e99cba1ee27fffe68afedef7c1dce40526f4d
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f0ef3fe9ebcf9f8f4fee3c065efe25035b81435ae8780b056e116dcb1a8a428f
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d6701719fda370188bcff5587df70fa5f1ce20ee16a7c8ee75ec8b4d555e1ea
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 2.8666666666666667,
5
  "eval_steps": 500,
6
- "global_step": 43000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -3033,6 +3033,41 @@
3033
  "learning_rate": 8.960000000000001e-07,
3034
  "loss": 0.0534,
3035
  "step": 43000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3036
  }
3037
  ],
3038
  "logging_steps": 100,
@@ -3052,7 +3087,7 @@
3052
  "attributes": {}
3053
  }
3054
  },
3055
- "total_flos": 1.0474075717632e+17,
3056
  "train_batch_size": 4,
3057
  "trial_name": null,
3058
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 2.9,
5
  "eval_steps": 500,
6
+ "global_step": 43500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
3033
  "learning_rate": 8.960000000000001e-07,
3034
  "loss": 0.0534,
3035
  "step": 43000
3036
+ },
3037
+ {
3038
+ "epoch": 2.873333333333333,
3039
+ "grad_norm": 0.22133620083332062,
3040
+ "learning_rate": 8.515555555555556e-07,
3041
+ "loss": 0.0489,
3042
+ "step": 43100
3043
+ },
3044
+ {
3045
+ "epoch": 2.88,
3046
+ "grad_norm": 0.23495450615882874,
3047
+ "learning_rate": 8.071111111111112e-07,
3048
+ "loss": 0.0524,
3049
+ "step": 43200
3050
+ },
3051
+ {
3052
+ "epoch": 2.8866666666666667,
3053
+ "grad_norm": 0.22712914645671844,
3054
+ "learning_rate": 7.626666666666667e-07,
3055
+ "loss": 0.0471,
3056
+ "step": 43300
3057
+ },
3058
+ {
3059
+ "epoch": 2.8933333333333335,
3060
+ "grad_norm": 0.3228939175605774,
3061
+ "learning_rate": 7.182222222222223e-07,
3062
+ "loss": 0.0495,
3063
+ "step": 43400
3064
+ },
3065
+ {
3066
+ "epoch": 2.9,
3067
+ "grad_norm": 0.20631560683250427,
3068
+ "learning_rate": 6.737777777777778e-07,
3069
+ "loss": 0.0521,
3070
+ "step": 43500
3071
  }
3072
  ],
3073
  "logging_steps": 100,
 
3087
  "attributes": {}
3088
  }
3089
  },
3090
+ "total_flos": 1.0595867295744e+17,
3091
  "train_batch_size": 4,
3092
  "trial_name": null,
3093
  "trial_params": null