masatochi commited on
Commit
7846c07
·
verified ·
1 Parent(s): ed8a6db

Training in progress, step 155, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a8079a1f6050a403017529fa32f132b84aa2e56a1cf9f664aee1753ce47829a4
3
  size 83945296
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42a3d9f5148756dc73c053314bcaa3e2db785a1e4a7d4376478c629737c24530
3
  size 83945296
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b2e4e0e76a4da37b7b356ce86b60b5081f75541abe08cb4688ae7d18700570f8
3
  size 43122580
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8fabf2ebf8b0f5f80e260e7eccd4debfda38d3d814c60433a54fdf5470ead0d
3
  size 43122580
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2d95327014bf0e992837899fdc63d1e3049ea8baca978228506724df9c716834
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1849ad9db14ff92103b9d227ede45e09b064472ac17db6af688392748c6c18b
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5a946cb282348d3fba8c242cd51f3b90b3dccbd24720ee6b6397a7e493e7b92c
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:382ece00c385c691790a218c7659825f5fdd28b4e63aa1032e7b069dd2944457
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.07335411699981662,
5
  "eval_steps": 34,
6
- "global_step": 150,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1097,6 +1097,41 @@
1097
  "learning_rate": 3.973653636207437e-05,
1098
  "loss": 1.0422,
1099
  "step": 150
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1100
  }
1101
  ],
1102
  "logging_steps": 1,
@@ -1116,7 +1151,7 @@
1116
  "attributes": {}
1117
  }
1118
  },
1119
- "total_flos": 6.658431293325312e+17,
1120
  "train_batch_size": 3,
1121
  "trial_name": null,
1122
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.07579925423314383,
5
  "eval_steps": 34,
6
+ "global_step": 155,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1097
  "learning_rate": 3.973653636207437e-05,
1098
  "loss": 1.0422,
1099
  "step": 150
1100
+ },
1101
+ {
1102
+ "epoch": 0.07384314444648206,
1103
+ "grad_norm": 1.3410860300064087,
1104
+ "learning_rate": 3.8272177871020723e-05,
1105
+ "loss": 1.0522,
1106
+ "step": 151
1107
+ },
1108
+ {
1109
+ "epoch": 0.0743321718931475,
1110
+ "grad_norm": 1.0031688213348389,
1111
+ "learning_rate": 3.682889937467493e-05,
1112
+ "loss": 0.9383,
1113
+ "step": 152
1114
+ },
1115
+ {
1116
+ "epoch": 0.07482119933981295,
1117
+ "grad_norm": 1.1532304286956787,
1118
+ "learning_rate": 3.540719375132129e-05,
1119
+ "loss": 1.0742,
1120
+ "step": 153
1121
+ },
1122
+ {
1123
+ "epoch": 0.07531022678647839,
1124
+ "grad_norm": 1.2422029972076416,
1125
+ "learning_rate": 3.400754651212776e-05,
1126
+ "loss": 1.0617,
1127
+ "step": 154
1128
+ },
1129
+ {
1130
+ "epoch": 0.07579925423314383,
1131
+ "grad_norm": 1.1498128175735474,
1132
+ "learning_rate": 3.263043563534428e-05,
1133
+ "loss": 1.1001,
1134
+ "step": 155
1135
  }
1136
  ],
1137
  "logging_steps": 1,
 
1151
  "attributes": {}
1152
  }
1153
  },
1154
+ "total_flos": 6.880379003102822e+17,
1155
  "train_batch_size": 3,
1156
  "trial_name": null,
1157
  "trial_params": null