rakhman-llm commited on
Commit
833ed50
·
verified ·
1 Parent(s): e7e57e8

Training in progress, step 42000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ddeba81fa20a347b7f4a8bf56911f542931a9c7a9a28e5b17d8415cffe0f05fc
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:745b8b1d7e5bd21d472de4928ff7acd2ffa83a677380359019a9c4799912b8d9
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c534f106883dfc1dc4725defbacef030c7dba5f5e0888cc081c7906d9d0b8453
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff26510f9fa790ef0a5e5f59b48491824122de83388e6b8dafee62d5541a9a5f
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3b282c8a79340622eb958db3d6110bd8f623dfad615139665f1498d8b00a6bed
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1538e2c63d940ef0e1b73a9e8d73309e15231b1a12e77d8cac6d2db2463579dc
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6abecd6b93d057c006d3ff371902f2a70b673a5fc3d8d5773b1c045cd3372a7b
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:89058407f5b943be63f8348b693549eede8380db14dc40de0b317aff021fa0d0
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 2.7666666666666666,
5
  "eval_steps": 500,
6
- "global_step": 41500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -2928,6 +2928,41 @@
2928
  "learning_rate": 1.5626666666666667e-06,
2929
  "loss": 0.0485,
2930
  "step": 41500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2931
  }
2932
  ],
2933
  "logging_steps": 100,
@@ -2947,7 +2982,7 @@
2947
  "attributes": {}
2948
  }
2949
  },
2950
- "total_flos": 1.0108700983296e+17,
2951
  "train_batch_size": 4,
2952
  "trial_name": null,
2953
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 2.8,
5
  "eval_steps": 500,
6
+ "global_step": 42000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
2928
  "learning_rate": 1.5626666666666667e-06,
2929
  "loss": 0.0485,
2930
  "step": 41500
2931
+ },
2932
+ {
2933
+ "epoch": 2.7733333333333334,
2934
+ "grad_norm": 0.2192191630601883,
2935
+ "learning_rate": 1.5182222222222223e-06,
2936
+ "loss": 0.0514,
2937
+ "step": 41600
2938
+ },
2939
+ {
2940
+ "epoch": 2.7800000000000002,
2941
+ "grad_norm": 0.17543815076351166,
2942
+ "learning_rate": 1.4737777777777778e-06,
2943
+ "loss": 0.0473,
2944
+ "step": 41700
2945
+ },
2946
+ {
2947
+ "epoch": 2.7866666666666666,
2948
+ "grad_norm": 0.25185197591781616,
2949
+ "learning_rate": 1.4293333333333334e-06,
2950
+ "loss": 0.0508,
2951
+ "step": 41800
2952
+ },
2953
+ {
2954
+ "epoch": 2.7933333333333334,
2955
+ "grad_norm": 0.22377026081085205,
2956
+ "learning_rate": 1.3848888888888889e-06,
2957
+ "loss": 0.0617,
2958
+ "step": 41900
2959
+ },
2960
+ {
2961
+ "epoch": 2.8,
2962
+ "grad_norm": 0.1886560320854187,
2963
+ "learning_rate": 1.3404444444444445e-06,
2964
+ "loss": 0.0504,
2965
+ "step": 42000
2966
  }
2967
  ],
2968
  "logging_steps": 100,
 
2982
  "attributes": {}
2983
  }
2984
  },
2985
+ "total_flos": 1.0230492561408e+17,
2986
  "train_batch_size": 4,
2987
  "trial_name": null,
2988
  "trial_params": null