masatochi commited on
Commit
76c3d2a
·
verified ·
1 Parent(s): da2ea7c

Training in progress, step 10, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aacc3e9459558a0715206f71408318788c1a138b86dbe25078d85b7744b159bd
3
  size 83945296
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55b792505145e876ad8a4936e10d257b699bbb0e93f0888613b910e08754b08b
3
  size 83945296
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9f60de47404974b8762b860dcaff847d20f15b263751923f8379db82721710a2
3
  size 43122580
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2458d479fc914e91e5bc04c2ea5b309f54fa4c1791211b07e0785347703eac8
3
  size 43122580
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5d11172069d3809e4b1d435d13824f7b7bfef9c55e762daff1ad8efa9119dcca
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1bf24cd4bf25e308928bff2034c8266e5fe111aaec9b6d5c14e665ac843e0bc
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:16178af669793c21eac8710459f1cc10bcf09af1984c5f2d80e76b26b329c8ba
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:67680b8b35ea230d53acee387012f8809825a2c511bd41340f42320f849e349d
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.0024451372333272204,
5
  "eval_steps": 34,
6
- "global_step": 5,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -50,6 +50,41 @@
50
  "learning_rate": 3.3333333333333335e-05,
51
  "loss": 3.6677,
52
  "step": 5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
53
  }
54
  ],
55
  "logging_steps": 1,
@@ -69,7 +104,7 @@
69
  "attributes": {}
70
  }
71
  },
72
- "total_flos": 2.219477097775104e+16,
73
  "train_batch_size": 3,
74
  "trial_name": null,
75
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.004890274466654441,
5
  "eval_steps": 34,
6
+ "global_step": 10,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
50
  "learning_rate": 3.3333333333333335e-05,
51
  "loss": 3.6677,
52
  "step": 5
53
+ },
54
+ {
55
+ "epoch": 0.0029341646799926646,
56
+ "grad_norm": 1.7147626876831055,
57
+ "learning_rate": 4e-05,
58
+ "loss": 3.8075,
59
+ "step": 6
60
+ },
61
+ {
62
+ "epoch": 0.0034231921266581087,
63
+ "grad_norm": 1.672584891319275,
64
+ "learning_rate": 4.666666666666667e-05,
65
+ "loss": 3.3499,
66
+ "step": 7
67
+ },
68
+ {
69
+ "epoch": 0.0039122195733235525,
70
+ "grad_norm": 2.0120997428894043,
71
+ "learning_rate": 5.333333333333333e-05,
72
+ "loss": 4.2653,
73
+ "step": 8
74
+ },
75
+ {
76
+ "epoch": 0.004401247019988997,
77
+ "grad_norm": 2.0601062774658203,
78
+ "learning_rate": 6e-05,
79
+ "loss": 3.3055,
80
+ "step": 9
81
+ },
82
+ {
83
+ "epoch": 0.004890274466654441,
84
+ "grad_norm": 2.0153276920318604,
85
+ "learning_rate": 6.666666666666667e-05,
86
+ "loss": 3.1547,
87
+ "step": 10
88
  }
89
  ],
90
  "logging_steps": 1,
 
104
  "attributes": {}
105
  }
106
  },
107
+ "total_flos": 4.438954195550208e+16,
108
  "train_batch_size": 3,
109
  "trial_name": null,
110
  "trial_params": null