End of training
Browse files- README.md +5 -5
- logs/events.out.tfevents.1707149028.4c55bcba7827.26.0 +3 -0
- model.safetensors +1 -1
- trainer_state.json +17 -3
README.md
CHANGED
@@ -13,12 +13,12 @@ should probably proofread and complete it, then remove this comment. -->
|
|
13 |
|
14 |
This model was trained from scratch on an unknown dataset.
|
15 |
It achieves the following results on the evaluation set:
|
16 |
-
- eval_loss: 0.
|
17 |
-
- eval_runtime: 1376.
|
18 |
-
- eval_samples_per_second: 108.
|
19 |
- eval_steps_per_second: 1.703
|
20 |
-
- epoch:
|
21 |
-
- step:
|
22 |
|
23 |
## Model description
|
24 |
|
|
|
13 |
|
14 |
This model was trained from scratch on an unknown dataset.
|
15 |
It achieves the following results on the evaluation set:
|
16 |
+
- eval_loss: 0.9588
|
17 |
+
- eval_runtime: 1376.5573
|
18 |
+
- eval_samples_per_second: 108.967
|
19 |
- eval_steps_per_second: 1.703
|
20 |
+
- epoch: 3.56
|
21 |
+
- step: 150000
|
22 |
|
23 |
## Model description
|
24 |
|
logs/events.out.tfevents.1707149028.4c55bcba7827.26.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:de511e2aff8aec50f56506a4d453b836ad0c2e858090c2a7f49c427fc8db70de
|
3 |
+
size 5154
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 497918592
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:569bc69dea165d5d6cfe12190fa1b33d3707f544b843c3a41ac3a46c44484a09
|
3 |
size 497918592
|
trainer_state.json
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
{
|
2 |
"best_metric": null,
|
3 |
"best_model_checkpoint": null,
|
4 |
-
"epoch":
|
5 |
"eval_steps": 50000,
|
6 |
-
"global_step":
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
@@ -35,6 +35,20 @@
|
|
35 |
"eval_samples_per_second": 108.974,
|
36 |
"eval_steps_per_second": 1.703,
|
37 |
"step": 100000
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
38 |
}
|
39 |
],
|
40 |
"logging_steps": 50000,
|
@@ -42,7 +56,7 @@
|
|
42 |
"num_input_tokens_seen": 0,
|
43 |
"num_train_epochs": 100,
|
44 |
"save_steps": 50000,
|
45 |
-
"total_flos":
|
46 |
"train_batch_size": 32,
|
47 |
"trial_name": null,
|
48 |
"trial_params": null
|
|
|
1 |
{
|
2 |
"best_metric": null,
|
3 |
"best_model_checkpoint": null,
|
4 |
+
"epoch": 3.5555134161372903,
|
5 |
"eval_steps": 50000,
|
6 |
+
"global_step": 150000,
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
|
|
35 |
"eval_samples_per_second": 108.974,
|
36 |
"eval_steps_per_second": 1.703,
|
37 |
"step": 100000
|
38 |
+
},
|
39 |
+
{
|
40 |
+
"epoch": 3.56,
|
41 |
+
"learning_rate": 4.940766331658292e-05,
|
42 |
+
"loss": 0.9833,
|
43 |
+
"step": 150000
|
44 |
+
},
|
45 |
+
{
|
46 |
+
"epoch": 3.56,
|
47 |
+
"eval_loss": 0.9588034152984619,
|
48 |
+
"eval_runtime": 1376.5573,
|
49 |
+
"eval_samples_per_second": 108.967,
|
50 |
+
"eval_steps_per_second": 1.703,
|
51 |
+
"step": 150000
|
52 |
}
|
53 |
],
|
54 |
"logging_steps": 50000,
|
|
|
56 |
"num_input_tokens_seen": 0,
|
57 |
"num_train_epochs": 100,
|
58 |
"save_steps": 50000,
|
59 |
+
"total_flos": 3.13547302895616e+17,
|
60 |
"train_batch_size": 32,
|
61 |
"trial_name": null,
|
62 |
"trial_params": null
|