WangXFng commited on
Commit
b933dc2
·
verified ·
1 Parent(s): ee47939

Model save

Browse files
README.md CHANGED
@@ -34,11 +34,11 @@ More information needed
34
 
35
  The following hyperparameters were used during training:
36
  - learning_rate: 0.0001
37
- - train_batch_size: 6
38
  - eval_batch_size: 8
39
  - seed: 42
40
  - gradient_accumulation_steps: 16
41
- - total_train_batch_size: 96
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: linear
44
  - lr_scheduler_warmup_steps: 2
 
34
 
35
  The following hyperparameters were used during training:
36
  - learning_rate: 0.0001
37
+ - train_batch_size: 16
38
  - eval_batch_size: 8
39
  - seed: 42
40
  - gradient_accumulation_steps: 16
41
+ - total_train_batch_size: 256
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: linear
44
  - lr_scheduler_warmup_steps: 2
adapter_config.json CHANGED
@@ -20,12 +20,12 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "k_proj",
24
- "down_proj",
25
  "q_proj",
 
 
 
26
  "o_proj",
27
  "v_proj",
28
- "up_proj",
29
  "gate_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
23
  "q_proj",
24
+ "down_proj",
25
+ "up_proj",
26
+ "k_proj",
27
  "o_proj",
28
  "v_proj",
 
29
  "gate_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:39ef7b5ed66ff345f822092e69d270531391d309f346e882bac6e246a0b98536
3
  size 1635969696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d209c54a5d95be6e1246f9254828bb85ebf2c9da9c8a5016d8aa65f0d84eb0ca
3
  size 1635969696
trainer_state.json CHANGED
@@ -1,172 +1,81 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 3.999089792017476,
5
  "eval_steps": 500,
6
- "global_step": 5492,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 0.18204159650480134,
13
- "grad_norm": 1.178935170173645,
14
- "learning_rate": 9.548269581056467e-05,
15
- "loss": 1.0653,
16
  "step": 250
17
  },
18
  {
19
- "epoch": 0.3640831930096027,
20
- "grad_norm": 1.0788908004760742,
21
- "learning_rate": 9.092896174863389e-05,
22
- "loss": 0.5664,
23
  "step": 500
24
  },
25
  {
26
- "epoch": 0.546124789514404,
27
- "grad_norm": 0.8843607902526855,
28
- "learning_rate": 8.63752276867031e-05,
29
- "loss": 0.4896,
30
  "step": 750
31
  },
32
  {
33
- "epoch": 0.7281663860192054,
34
- "grad_norm": 0.8972049951553345,
35
- "learning_rate": 8.182149362477231e-05,
36
- "loss": 0.4628,
37
  "step": 1000
38
  },
39
  {
40
- "epoch": 0.9102079825240067,
41
- "grad_norm": 0.7201969027519226,
42
- "learning_rate": 7.726775956284153e-05,
43
- "loss": 0.4461,
44
  "step": 1250
45
  },
46
  {
47
- "epoch": 1.092249579028808,
48
- "grad_norm": 0.7767462134361267,
49
- "learning_rate": 7.271402550091076e-05,
50
- "loss": 0.4344,
51
  "step": 1500
52
  },
53
  {
54
- "epoch": 1.2742911755336095,
55
- "grad_norm": 0.7544079422950745,
56
- "learning_rate": 6.816029143897996e-05,
57
- "loss": 0.425,
58
  "step": 1750
59
  },
60
  {
61
- "epoch": 1.4563327720384107,
62
- "grad_norm": 0.7420190572738647,
63
- "learning_rate": 6.360655737704918e-05,
64
- "loss": 0.4211,
65
  "step": 2000
66
  },
67
  {
68
- "epoch": 1.6383743685432122,
69
- "grad_norm": 0.8790698051452637,
70
- "learning_rate": 5.90528233151184e-05,
71
- "loss": 0.4094,
72
- "step": 2250
73
- },
74
- {
75
- "epoch": 1.8204159650480136,
76
- "grad_norm": 0.8602856397628784,
77
- "learning_rate": 5.449908925318762e-05,
78
- "loss": 0.3953,
79
- "step": 2500
80
- },
81
- {
82
- "epoch": 2.0024575615528146,
83
- "grad_norm": 0.9606088995933533,
84
- "learning_rate": 4.994535519125683e-05,
85
- "loss": 0.3816,
86
- "step": 2750
87
- },
88
- {
89
- "epoch": 2.184499158057616,
90
- "grad_norm": 1.0402580499649048,
91
- "learning_rate": 4.539162112932605e-05,
92
- "loss": 0.3574,
93
- "step": 3000
94
- },
95
- {
96
- "epoch": 2.3665407545624175,
97
- "grad_norm": 1.0433388948440552,
98
- "learning_rate": 4.083788706739526e-05,
99
- "loss": 0.3438,
100
- "step": 3250
101
- },
102
- {
103
- "epoch": 2.548582351067219,
104
- "grad_norm": 1.0822595357894897,
105
- "learning_rate": 3.6284153005464486e-05,
106
- "loss": 0.326,
107
- "step": 3500
108
- },
109
- {
110
- "epoch": 2.7306239475720204,
111
- "grad_norm": 1.1437020301818848,
112
- "learning_rate": 3.17304189435337e-05,
113
- "loss": 0.3155,
114
- "step": 3750
115
- },
116
- {
117
- "epoch": 2.9126655440768214,
118
- "grad_norm": 1.1588534116744995,
119
- "learning_rate": 2.7176684881602916e-05,
120
- "loss": 0.3043,
121
- "step": 4000
122
- },
123
- {
124
- "epoch": 3.094707140581623,
125
- "grad_norm": 1.0676839351654053,
126
- "learning_rate": 2.262295081967213e-05,
127
- "loss": 0.2888,
128
- "step": 4250
129
- },
130
- {
131
- "epoch": 3.2767487370864243,
132
- "grad_norm": 1.149923324584961,
133
- "learning_rate": 1.806921675774135e-05,
134
- "loss": 0.2762,
135
- "step": 4500
136
- },
137
- {
138
- "epoch": 3.458790333591226,
139
- "grad_norm": 1.1598175764083862,
140
- "learning_rate": 1.3515482695810567e-05,
141
- "loss": 0.2721,
142
- "step": 4750
143
- },
144
- {
145
- "epoch": 3.640831930096027,
146
- "grad_norm": 1.1374166011810303,
147
- "learning_rate": 8.961748633879782e-06,
148
- "loss": 0.2682,
149
- "step": 5000
150
- },
151
- {
152
- "epoch": 3.8228735266008282,
153
- "grad_norm": 1.123226523399353,
154
- "learning_rate": 4.408014571948998e-06,
155
- "loss": 0.2637,
156
- "step": 5250
157
- },
158
- {
159
- "epoch": 3.999089792017476,
160
- "step": 5492,
161
- "total_flos": 1.2110720614916813e+18,
162
- "train_loss": 0.3990837633392712,
163
- "train_runtime": 19256.2222,
164
- "train_samples_per_second": 27.386,
165
- "train_steps_per_second": 0.285
166
  }
167
  ],
168
  "logging_steps": 250,
169
- "max_steps": 5492,
170
  "num_input_tokens_seen": 0,
171
  "num_train_epochs": 4,
172
  "save_steps": 500,
@@ -182,8 +91,8 @@
182
  "attributes": {}
183
  }
184
  },
185
- "total_flos": 1.2110720614916813e+18,
186
- "train_batch_size": 6,
187
  "trial_name": null,
188
  "trial_params": null
189
  }
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 4.0,
5
  "eval_steps": 500,
6
+ "global_step": 2060,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 0.4854368932038835,
13
+ "grad_norm": 0.5123344659805298,
14
+ "learning_rate": 8.794946550048592e-05,
15
+ "loss": 0.9496,
16
  "step": 250
17
  },
18
  {
19
+ "epoch": 0.970873786407767,
20
+ "grad_norm": 0.8029014468193054,
21
+ "learning_rate": 7.580174927113704e-05,
22
+ "loss": 0.6426,
23
  "step": 500
24
  },
25
  {
26
+ "epoch": 1.4563106796116505,
27
+ "grad_norm": 0.9907676577568054,
28
+ "learning_rate": 6.365403304178815e-05,
29
+ "loss": 0.5632,
30
  "step": 750
31
  },
32
  {
33
+ "epoch": 1.941747572815534,
34
+ "grad_norm": 0.9990822672843933,
35
+ "learning_rate": 5.150631681243926e-05,
36
+ "loss": 0.5117,
37
  "step": 1000
38
  },
39
  {
40
+ "epoch": 2.4271844660194173,
41
+ "grad_norm": 0.7601630687713623,
42
+ "learning_rate": 3.9358600583090386e-05,
43
+ "loss": 0.4772,
44
  "step": 1250
45
  },
46
  {
47
+ "epoch": 2.912621359223301,
48
+ "grad_norm": 0.7860077023506165,
49
+ "learning_rate": 2.72108843537415e-05,
50
+ "loss": 0.4564,
51
  "step": 1500
52
  },
53
  {
54
+ "epoch": 3.3980582524271843,
55
+ "grad_norm": 0.5943681597709656,
56
+ "learning_rate": 1.5063168124392615e-05,
57
+ "loss": 0.4392,
58
  "step": 1750
59
  },
60
  {
61
+ "epoch": 3.883495145631068,
62
+ "grad_norm": 0.5319840908050537,
63
+ "learning_rate": 2.915451895043732e-06,
64
+ "loss": 0.4318,
65
  "step": 2000
66
  },
67
  {
68
+ "epoch": 4.0,
69
+ "step": 2060,
70
+ "total_flos": 1.4445804612483994e+18,
71
+ "train_loss": 0.5552371376926459,
72
+ "train_runtime": 17453.3458,
73
+ "train_samples_per_second": 30.215,
74
+ "train_steps_per_second": 0.118
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
75
  }
76
  ],
77
  "logging_steps": 250,
78
+ "max_steps": 2060,
79
  "num_input_tokens_seen": 0,
80
  "num_train_epochs": 4,
81
  "save_steps": 500,
 
91
  "attributes": {}
92
  }
93
  },
94
+ "total_flos": 1.4445804612483994e+18,
95
+ "train_batch_size": 16,
96
  "trial_name": null,
97
  "trial_params": null
98
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c43f3ec46e656601dba5c8176f6da304521a9b5a664289cfb437165d2d3c7259
3
  size 5240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:617a99532cd91765e0705e15346b67e9640bce2631da172c620125a4f5bbf68b
3
  size 5240