masatochi commited on
Commit
9372655
·
verified ·
1 Parent(s): f9f6fb7

Training in progress, step 125, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dfa8c53e5e820044af6c50ee1224154243471f291bb2ea626ff5f3dbb284aa50
3
  size 83945296
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1628b2bb2347dfb854cd6be60adf8d3e2e41848076d02d48eb2cbf0c189fdd5d
3
  size 83945296
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f0381c9373d63945244b6821e06cbd4bd417ba7642a79644335119bda023a7c7
3
  size 43122580
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f42869777e41217dd1ac60269c980858e1e77e66d0f40c379fc76a41f09e260
3
  size 43122580
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5b37c228faac27493d5dc2700eeaf67fe98afdbd0e55a68eaf314f21f0aea103
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78ae4de6cae258994b2e610a17502d0e0897089893ff3ea440cdda7e7a8e9774
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ba16893b2c4735e8eaf86592331a8dda9b3bcccecd302e828000513277487239
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b577939e8ae09a93269bdd1ffbcc4ef41ec4027476aa914ab19034c5a6ebf492
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.05868329359985329,
5
  "eval_steps": 34,
6
- "global_step": 120,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -879,6 +879,41 @@
879
  "learning_rate": 9.077316405366981e-05,
880
  "loss": 0.9481,
881
  "step": 120
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
882
  }
883
  ],
884
  "logging_steps": 1,
@@ -898,7 +933,7 @@
898
  "attributes": {}
899
  }
900
  },
901
- "total_flos": 5.3267450346602496e+17,
902
  "train_batch_size": 3,
903
  "trial_name": null,
904
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.061128430833180515,
5
  "eval_steps": 34,
6
+ "global_step": 125,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
879
  "learning_rate": 9.077316405366981e-05,
880
  "loss": 0.9481,
881
  "step": 120
882
+ },
883
+ {
884
+ "epoch": 0.05917232104651873,
885
+ "grad_norm": 1.3437144756317139,
886
+ "learning_rate": 8.893473181084994e-05,
887
+ "loss": 1.1713,
888
+ "step": 121
889
+ },
890
+ {
891
+ "epoch": 0.05966134849318418,
892
+ "grad_norm": 1.3384225368499756,
893
+ "learning_rate": 8.710007834697969e-05,
894
+ "loss": 1.0073,
895
+ "step": 122
896
+ },
897
+ {
898
+ "epoch": 0.06015037593984962,
899
+ "grad_norm": 1.2030497789382935,
900
+ "learning_rate": 8.526983019453623e-05,
901
+ "loss": 1.2027,
902
+ "step": 123
903
+ },
904
+ {
905
+ "epoch": 0.06063940338651507,
906
+ "grad_norm": 1.4308159351348877,
907
+ "learning_rate": 8.344461238158699e-05,
908
+ "loss": 1.0608,
909
+ "step": 124
910
+ },
911
+ {
912
+ "epoch": 0.061128430833180515,
913
+ "grad_norm": 1.2718294858932495,
914
+ "learning_rate": 8.162504821834295e-05,
915
+ "loss": 1.0317,
916
+ "step": 125
917
  }
918
  ],
919
  "logging_steps": 1,
 
933
  "attributes": {}
934
  }
935
  },
936
+ "total_flos": 5.54869274443776e+17,
937
  "train_batch_size": 3,
938
  "trial_name": null,
939
  "trial_params": null