rakhman-llm commited on
Commit
9111139
·
verified ·
1 Parent(s): 9fba5d2

Training in progress, step 30000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:66cae8324741bfb1c3cbb4cfed2559c012d8b417990bdb0c6e1b400f774dcc4e
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8367a24eb06a34cbf04ec05d49423cacadb43f1fbc7c9b46cd3bc8f7d202aeea
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c9055ac327c15e228dd6a4caa71093d65f0ca825571af69e5994cc7332daa9be
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69cfb99a1d47731eb0bd0f0aa8008f24af3b21461df86591e07ee79decd04b36
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8eba248c4c107e3630aeadfd11e69cd930737cf324d325de259160a726c9e173
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59297a887ee86d9ae2f8461a820add9f102e062ed6b208e7b0f014f86659cc1d
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b2b394b8045eb22d77503865620f9a9d287ea2ba93ffd3839a0365feff6f4d8b
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ece32524ff12cd47152c7073cf4a24f61bf2d3f6e989a1d942aaad706c29e5e
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 1.9666666666666668,
5
  "eval_steps": 500,
6
- "global_step": 29500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -2080,6 +2080,41 @@
2080
  "learning_rate": 6.893333333333334e-06,
2081
  "loss": 0.0545,
2082
  "step": 29500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2083
  }
2084
  ],
2085
  "logging_steps": 100,
@@ -2099,7 +2134,7 @@
2099
  "attributes": {}
2100
  }
2101
  },
2102
- "total_flos": 7.185703108608e+16,
2103
  "train_batch_size": 4,
2104
  "trial_name": null,
2105
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 2.0,
5
  "eval_steps": 500,
6
+ "global_step": 30000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
2080
  "learning_rate": 6.893333333333334e-06,
2081
  "loss": 0.0545,
2082
  "step": 29500
2083
+ },
2084
+ {
2085
+ "epoch": 1.9733333333333334,
2086
+ "grad_norm": 0.3285749852657318,
2087
+ "learning_rate": 6.849333333333333e-06,
2088
+ "loss": 0.0566,
2089
+ "step": 29600
2090
+ },
2091
+ {
2092
+ "epoch": 1.98,
2093
+ "grad_norm": 0.2452673316001892,
2094
+ "learning_rate": 6.80488888888889e-06,
2095
+ "loss": 0.0501,
2096
+ "step": 29700
2097
+ },
2098
+ {
2099
+ "epoch": 1.9866666666666668,
2100
+ "grad_norm": 0.06861041486263275,
2101
+ "learning_rate": 6.760444444444445e-06,
2102
+ "loss": 0.0554,
2103
+ "step": 29800
2104
+ },
2105
+ {
2106
+ "epoch": 1.9933333333333332,
2107
+ "grad_norm": 0.2544171214103699,
2108
+ "learning_rate": 6.716000000000001e-06,
2109
+ "loss": 0.058,
2110
+ "step": 29900
2111
+ },
2112
+ {
2113
+ "epoch": 2.0,
2114
+ "grad_norm": 0.22868148982524872,
2115
+ "learning_rate": 6.671555555555556e-06,
2116
+ "loss": 0.0569,
2117
+ "step": 30000
2118
  }
2119
  ],
2120
  "logging_steps": 100,
 
2134
  "attributes": {}
2135
  }
2136
  },
2137
+ "total_flos": 7.30749468672e+16,
2138
  "train_batch_size": 4,
2139
  "trial_name": null,
2140
  "trial_params": null