ghost613 commited on
Commit
6f2e634
·
verified ·
1 Parent(s): 74fda2d

Upload 4 files

Browse files
adapter_config.json CHANGED
@@ -15,17 +15,17 @@
15
  "megatron_core": "megatron.core",
16
  "modules_to_save": null,
17
  "peft_type": "LORA",
18
- "r": 32,
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
- "up_proj",
23
- "k_proj",
24
- "q_proj",
25
- "gate_proj",
26
  "down_proj",
 
 
27
  "v_proj",
28
- "o_proj"
 
 
29
  ],
30
  "task_type": "CAUSAL_LM",
31
  "use_rslora": false
 
15
  "megatron_core": "megatron.core",
16
  "modules_to_save": null,
17
  "peft_type": "LORA",
18
+ "r": 16,
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
 
 
 
 
22
  "down_proj",
23
+ "gate_proj",
24
+ "o_proj",
25
  "v_proj",
26
+ "k_proj",
27
+ "q_proj",
28
+ "up_proj"
29
  ],
30
  "task_type": "CAUSAL_LM",
31
  "use_rslora": false
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a9ce050e4e4496519431daaf68b3ed0aebfbb4002facbbaf7ae1da64ad590a53
3
- size 156926880
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f81d394abe35a4ee6127ee7258b849a9a14cc2614dd80c4370b456ce23b0bda
3
+ size 78480072
trainer_state.json ADDED
@@ -0,0 +1,186 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.6502026915550232,
3
+ "best_model_checkpoint": "gemma2_on_korean_summary/checkpoint-220",
4
+ "epoch": 2.8947368421052633,
5
+ "eval_steps": 20,
6
+ "global_step": 220,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.26,
13
+ "grad_norm": 0.9722719192504883,
14
+ "learning_rate": 2e-05,
15
+ "loss": 1.1522,
16
+ "step": 20
17
+ },
18
+ {
19
+ "epoch": 0.26,
20
+ "eval_loss": 1.033677101135254,
21
+ "eval_runtime": 209.4976,
22
+ "eval_samples_per_second": 0.955,
23
+ "eval_steps_per_second": 0.477,
24
+ "step": 20
25
+ },
26
+ {
27
+ "epoch": 0.53,
28
+ "grad_norm": 1.436105489730835,
29
+ "learning_rate": 4e-05,
30
+ "loss": 0.9522,
31
+ "step": 40
32
+ },
33
+ {
34
+ "epoch": 0.53,
35
+ "eval_loss": 0.8600569367408752,
36
+ "eval_runtime": 206.2633,
37
+ "eval_samples_per_second": 0.97,
38
+ "eval_steps_per_second": 0.485,
39
+ "step": 40
40
+ },
41
+ {
42
+ "epoch": 0.79,
43
+ "grad_norm": 1.6157077550888062,
44
+ "learning_rate": 4.933333333333334e-05,
45
+ "loss": 0.8405,
46
+ "step": 60
47
+ },
48
+ {
49
+ "epoch": 0.79,
50
+ "eval_loss": 0.7875047326087952,
51
+ "eval_runtime": 207.822,
52
+ "eval_samples_per_second": 0.962,
53
+ "eval_steps_per_second": 0.481,
54
+ "step": 60
55
+ },
56
+ {
57
+ "epoch": 1.05,
58
+ "grad_norm": 1.4430443048477173,
59
+ "learning_rate": 4.8e-05,
60
+ "loss": 0.7687,
61
+ "step": 80
62
+ },
63
+ {
64
+ "epoch": 1.05,
65
+ "eval_loss": 0.7403695583343506,
66
+ "eval_runtime": 210.2054,
67
+ "eval_samples_per_second": 0.951,
68
+ "eval_steps_per_second": 0.476,
69
+ "step": 80
70
+ },
71
+ {
72
+ "epoch": 1.32,
73
+ "grad_norm": 1.640892505645752,
74
+ "learning_rate": 4.666666666666667e-05,
75
+ "loss": 0.6901,
76
+ "step": 100
77
+ },
78
+ {
79
+ "epoch": 1.32,
80
+ "eval_loss": 0.7110423445701599,
81
+ "eval_runtime": 206.6708,
82
+ "eval_samples_per_second": 0.968,
83
+ "eval_steps_per_second": 0.484,
84
+ "step": 100
85
+ },
86
+ {
87
+ "epoch": 1.58,
88
+ "grad_norm": 1.6705243587493896,
89
+ "learning_rate": 4.5333333333333335e-05,
90
+ "loss": 0.6588,
91
+ "step": 120
92
+ },
93
+ {
94
+ "epoch": 1.58,
95
+ "eval_loss": 0.6947363018989563,
96
+ "eval_runtime": 208.6184,
97
+ "eval_samples_per_second": 0.959,
98
+ "eval_steps_per_second": 0.479,
99
+ "step": 120
100
+ },
101
+ {
102
+ "epoch": 1.84,
103
+ "grad_norm": 1.7401701211929321,
104
+ "learning_rate": 4.4000000000000006e-05,
105
+ "loss": 0.6698,
106
+ "step": 140
107
+ },
108
+ {
109
+ "epoch": 1.84,
110
+ "eval_loss": 0.6739406585693359,
111
+ "eval_runtime": 207.3092,
112
+ "eval_samples_per_second": 0.965,
113
+ "eval_steps_per_second": 0.482,
114
+ "step": 140
115
+ },
116
+ {
117
+ "epoch": 2.11,
118
+ "grad_norm": 1.804783582687378,
119
+ "learning_rate": 4.266666666666667e-05,
120
+ "loss": 0.6075,
121
+ "step": 160
122
+ },
123
+ {
124
+ "epoch": 2.11,
125
+ "eval_loss": 0.6695161461830139,
126
+ "eval_runtime": 206.5895,
127
+ "eval_samples_per_second": 0.968,
128
+ "eval_steps_per_second": 0.484,
129
+ "step": 160
130
+ },
131
+ {
132
+ "epoch": 2.37,
133
+ "grad_norm": 2.2475624084472656,
134
+ "learning_rate": 4.133333333333333e-05,
135
+ "loss": 0.5221,
136
+ "step": 180
137
+ },
138
+ {
139
+ "epoch": 2.37,
140
+ "eval_loss": 0.6713693141937256,
141
+ "eval_runtime": 209.2432,
142
+ "eval_samples_per_second": 0.956,
143
+ "eval_steps_per_second": 0.478,
144
+ "step": 180
145
+ },
146
+ {
147
+ "epoch": 2.63,
148
+ "grad_norm": 2.268982172012329,
149
+ "learning_rate": 4e-05,
150
+ "loss": 0.4865,
151
+ "step": 200
152
+ },
153
+ {
154
+ "epoch": 2.63,
155
+ "eval_loss": 0.6660403609275818,
156
+ "eval_runtime": 205.9824,
157
+ "eval_samples_per_second": 0.971,
158
+ "eval_steps_per_second": 0.485,
159
+ "step": 200
160
+ },
161
+ {
162
+ "epoch": 2.89,
163
+ "grad_norm": 2.3548617362976074,
164
+ "learning_rate": 3.866666666666667e-05,
165
+ "loss": 0.5239,
166
+ "step": 220
167
+ },
168
+ {
169
+ "epoch": 2.89,
170
+ "eval_loss": 0.6502026915550232,
171
+ "eval_runtime": 206.4589,
172
+ "eval_samples_per_second": 0.969,
173
+ "eval_steps_per_second": 0.484,
174
+ "step": 220
175
+ }
176
+ ],
177
+ "logging_steps": 20,
178
+ "max_steps": 800,
179
+ "num_input_tokens_seen": 0,
180
+ "num_train_epochs": 11,
181
+ "save_steps": 20,
182
+ "total_flos": 1.917650992083763e+16,
183
+ "train_batch_size": 2,
184
+ "trial_name": null,
185
+ "trial_params": null
186
+ }