rakhman-llm commited on
Commit
e8c557f
·
verified ·
1 Parent(s): 719d32e

Training in progress, step 1000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3331e477755949427e95b7dca968d6817d55bd6f917b4a8c3697b379a8dc4de9
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91ffd4788c4ea759629dc808cb051ca7eab384e5fb8ae9dfa2d14c415beef6ab
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d149c0078e7e5d6a67271171e08cca30c493c0a67019446d7d12b6b87d869ef3
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:282c89d11dc548c714ac38e7fb510faacb72a5ed4d68eccfbd51cd63307f0bdc
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:eeb00119856714d9031cbc8ce0f99dda5a9e485f405270a39f2ed5786c90b1b6
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a7bccffbe4567a72e1b938cd6f7b1d19b05b0d573111e28fd70761fd6902354
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0e8e1fdd83e33f36b55c6b8b26ed0abf72653733f5da9946c6d0ff99fd8c1f8d
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86c73fa767afb78abc2698cb0a4494c23586bafd962eae88fa8af29f90657dc4
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.03333333333333333,
5
  "eval_steps": 500,
6
- "global_step": 500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -42,6 +42,41 @@
42
  "learning_rate": 1.9778666666666667e-05,
43
  "loss": 0.0887,
44
  "step": 500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
45
  }
46
  ],
47
  "logging_steps": 100,
@@ -61,7 +96,7 @@
61
  "attributes": {}
62
  }
63
  },
64
- "total_flos": 1217915781120000.0,
65
  "train_batch_size": 4,
66
  "trial_name": null,
67
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.06666666666666667,
5
  "eval_steps": 500,
6
+ "global_step": 1000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
42
  "learning_rate": 1.9778666666666667e-05,
43
  "loss": 0.0887,
44
  "step": 500
45
+ },
46
+ {
47
+ "epoch": 0.04,
48
+ "grad_norm": 0.30884480476379395,
49
+ "learning_rate": 1.9734222222222223e-05,
50
+ "loss": 0.0926,
51
+ "step": 600
52
+ },
53
+ {
54
+ "epoch": 0.04666666666666667,
55
+ "grad_norm": 0.14888174831867218,
56
+ "learning_rate": 1.9689777777777778e-05,
57
+ "loss": 0.0859,
58
+ "step": 700
59
+ },
60
+ {
61
+ "epoch": 0.05333333333333334,
62
+ "grad_norm": 0.23004640638828278,
63
+ "learning_rate": 1.9645333333333333e-05,
64
+ "loss": 0.0851,
65
+ "step": 800
66
+ },
67
+ {
68
+ "epoch": 0.06,
69
+ "grad_norm": 0.38090068101882935,
70
+ "learning_rate": 1.9600888888888892e-05,
71
+ "loss": 0.0824,
72
+ "step": 900
73
+ },
74
+ {
75
+ "epoch": 0.06666666666666667,
76
+ "grad_norm": 0.12699204683303833,
77
+ "learning_rate": 1.9556444444444447e-05,
78
+ "loss": 0.0851,
79
+ "step": 1000
80
  }
81
  ],
82
  "logging_steps": 100,
 
96
  "attributes": {}
97
  }
98
  },
99
+ "total_flos": 2435831562240000.0,
100
  "train_batch_size": 4,
101
  "trial_name": null,
102
  "trial_params": null
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5da3872b72c31dc28c756543e89c64090a6cb637a3a5de22e6dfe587939acf42
3
  size 5432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:20d7bfd722c356d3138649856f039df4c593cfd472f7e9aa290a4571f1e6327f
3
  size 5432