Step... (26000/50000 | Loss: 1.6362030506134033, Acc: 0.6691190600395203): 53%|█████████████▊ | 26656/50000 [10:36:57<11:06:38, 1.71s/it]
Browse files- flax_model.msgpack +1 -1
- outputs/checkpoints/checkpoint-20000/training_state.json +0 -1
- outputs/checkpoints/checkpoint-21000/training_state.json +0 -1
- outputs/checkpoints/{checkpoint-20000 → checkpoint-25000}/config.json +0 -0
- outputs/checkpoints/{checkpoint-20000 → checkpoint-25000}/data_collator.joblib +0 -0
- outputs/checkpoints/{checkpoint-20000 → checkpoint-25000}/flax_model.msgpack +1 -1
- outputs/checkpoints/{checkpoint-20000 → checkpoint-25000}/optimizer_state.msgpack +1 -1
- outputs/checkpoints/{checkpoint-20000 → checkpoint-25000}/training_args.joblib +0 -0
- outputs/checkpoints/checkpoint-25000/training_state.json +1 -0
- outputs/checkpoints/{checkpoint-21000 → checkpoint-26000}/config.json +0 -0
- outputs/checkpoints/{checkpoint-21000 → checkpoint-26000}/data_collator.joblib +0 -0
- outputs/checkpoints/{checkpoint-21000 → checkpoint-26000}/flax_model.msgpack +1 -1
- outputs/checkpoints/{checkpoint-21000 → checkpoint-26000}/optimizer_state.msgpack +1 -1
- outputs/checkpoints/{checkpoint-21000 → checkpoint-26000}/training_args.joblib +0 -0
- outputs/checkpoints/checkpoint-26000/training_state.json +1 -0
- outputs/events.out.tfevents.1627258355.tablespoon.3000110.3.v2 +2 -2
- outputs/flax_model.msgpack +1 -1
- outputs/optimizer_state.msgpack +1 -1
- outputs/training_state.json +1 -1
- pytorch_model.bin +1 -1
- run_stream.512.log +0 -0
- wandb/run-20210726_001233-17u6inbn/files/output.log +1704 -0
- wandb/run-20210726_001233-17u6inbn/files/wandb-summary.json +1 -1
- wandb/run-20210726_001233-17u6inbn/logs/debug-internal.log +2 -2
- wandb/run-20210726_001233-17u6inbn/run-17u6inbn.wandb +2 -2
flax_model.msgpack
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 249750019
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5ceb7903825a5802361e2abfa48cc22549619333ade477fa0502eca94fff7518
|
3 |
size 249750019
|
outputs/checkpoints/checkpoint-20000/training_state.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"step": 20001}
|
|
|
|
outputs/checkpoints/checkpoint-21000/training_state.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"step": 21001}
|
|
|
|
outputs/checkpoints/{checkpoint-20000 → checkpoint-25000}/config.json
RENAMED
File without changes
|
outputs/checkpoints/{checkpoint-20000 → checkpoint-25000}/data_collator.joblib
RENAMED
File without changes
|
outputs/checkpoints/{checkpoint-20000 → checkpoint-25000}/flax_model.msgpack
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 249750019
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3f01f7bd9dc37008fe586042d0858630facb5e28a5aabe227d941a460b9dac62
|
3 |
size 249750019
|
outputs/checkpoints/{checkpoint-20000 → checkpoint-25000}/optimizer_state.msgpack
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 499500278
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fd4e308d24cbf75134aac7314daf08bd6739fe89d8e0c1dc9de2428e0a4a588a
|
3 |
size 499500278
|
outputs/checkpoints/{checkpoint-20000 → checkpoint-25000}/training_args.joblib
RENAMED
File without changes
|
outputs/checkpoints/checkpoint-25000/training_state.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"step": 25001}
|
outputs/checkpoints/{checkpoint-21000 → checkpoint-26000}/config.json
RENAMED
File without changes
|
outputs/checkpoints/{checkpoint-21000 → checkpoint-26000}/data_collator.joblib
RENAMED
File without changes
|
outputs/checkpoints/{checkpoint-21000 → checkpoint-26000}/flax_model.msgpack
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 249750019
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5ceb7903825a5802361e2abfa48cc22549619333ade477fa0502eca94fff7518
|
3 |
size 249750019
|
outputs/checkpoints/{checkpoint-21000 → checkpoint-26000}/optimizer_state.msgpack
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 499500278
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:22861588524628232613a207d4ef751ae2280267840c912f9817827bd0747a14
|
3 |
size 499500278
|
outputs/checkpoints/{checkpoint-21000 → checkpoint-26000}/training_args.joblib
RENAMED
File without changes
|
outputs/checkpoints/checkpoint-26000/training_state.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"step": 26001}
|
outputs/events.out.tfevents.1627258355.tablespoon.3000110.3.v2
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c07c9e738eed842b50ab4e82610a59b50a7f87c6b9aeeff615946beccf6da7a2
|
3 |
+
size 3922999
|
outputs/flax_model.msgpack
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 249750019
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5ceb7903825a5802361e2abfa48cc22549619333ade477fa0502eca94fff7518
|
3 |
size 249750019
|
outputs/optimizer_state.msgpack
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 499500278
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:22861588524628232613a207d4ef751ae2280267840c912f9817827bd0747a14
|
3 |
size 499500278
|
outputs/training_state.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"step":
|
|
|
1 |
+
{"step": 26001}
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 498858859
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:17b4d4ba780e464164915cfe3081ea36a32c7f31487d8cc141e99fe4a4e44171
|
3 |
size 498858859
|
run_stream.512.log
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
wandb/run-20210726_001233-17u6inbn/files/output.log
CHANGED
@@ -16347,6 +16347,1710 @@ You should probably TRAIN this model on a down-stream task to be able to use it
|
|
16347 |
|
16348 |
|
16349 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
16350 |
|
16351 |
|
16352 |
|
|
|
16347 |
|
16348 |
|
16349 |
|
16350 |
+
|
16351 |
+
|
16352 |
+
|
16353 |
+
|
16354 |
+
|
16355 |
+
|
16356 |
+
|
16357 |
+
|
16358 |
+
|
16359 |
+
|
16360 |
+
|
16361 |
+
|
16362 |
+
|
16363 |
+
|
16364 |
+
|
16365 |
+
|
16366 |
+
|
16367 |
+
|
16368 |
+
|
16369 |
+
|
16370 |
+
|
16371 |
+
|
16372 |
+
|
16373 |
+
|
16374 |
+
|
16375 |
+
|
16376 |
+
|
16377 |
+
|
16378 |
+
|
16379 |
+
|
16380 |
+
|
16381 |
+
|
16382 |
+
|
16383 |
+
|
16384 |
+
|
16385 |
+
|
16386 |
+
|
16387 |
+
|
16388 |
+
|
16389 |
+
|
16390 |
+
|
16391 |
+
|
16392 |
+
|
16393 |
+
|
16394 |
+
|
16395 |
+
|
16396 |
+
|
16397 |
+
|
16398 |
+
|
16399 |
+
|
16400 |
+
|
16401 |
+
|
16402 |
+
|
16403 |
+
|
16404 |
+
|
16405 |
+
|
16406 |
+
|
16407 |
+
|
16408 |
+
|
16409 |
+
|
16410 |
+
|
16411 |
+
|
16412 |
+
|
16413 |
+
|
16414 |
+
|
16415 |
+
|
16416 |
+
|
16417 |
+
|
16418 |
+
|
16419 |
+
|
16420 |
+
|
16421 |
+
|
16422 |
+
|
16423 |
+
|
16424 |
+
|
16425 |
+
|
16426 |
+
|
16427 |
+
|
16428 |
+
|
16429 |
+
|
16430 |
+
|
16431 |
+
|
16432 |
+
|
16433 |
+
|
16434 |
+
|
16435 |
+
|
16436 |
+
|
16437 |
+
|
16438 |
+
|
16439 |
+
|
16440 |
+
|
16441 |
+
|
16442 |
+
|
16443 |
+
|
16444 |
+
|
16445 |
+
|
16446 |
+
|
16447 |
+
|
16448 |
+
|
16449 |
+
|
16450 |
+
|
16451 |
+
|
16452 |
+
|
16453 |
+
|
16454 |
+
|
16455 |
+
|
16456 |
+
|
16457 |
+
|
16458 |
+
|
16459 |
+
|
16460 |
+
|
16461 |
+
|
16462 |
+
|
16463 |
+
|
16464 |
+
|
16465 |
+
|
16466 |
+
|
16467 |
+
|
16468 |
+
|
16469 |
+
|
16470 |
+
|
16471 |
+
|
16472 |
+
|
16473 |
+
|
16474 |
+
|
16475 |
+
|
16476 |
+
|
16477 |
+
|
16478 |
+
|
16479 |
+
|
16480 |
+
|
16481 |
+
|
16482 |
+
|
16483 |
+
|
16484 |
+
|
16485 |
+
|
16486 |
+
|
16487 |
+
|
16488 |
+
|
16489 |
+
|
16490 |
+
|
16491 |
+
|
16492 |
+
|
16493 |
+
|
16494 |
+
|
16495 |
+
|
16496 |
+
|
16497 |
+
|
16498 |
+
|
16499 |
+
|
16500 |
+
|
16501 |
+
|
16502 |
+
|
16503 |
+
|
16504 |
+
|
16505 |
+
|
16506 |
+
|
16507 |
+
|
16508 |
+
|
16509 |
+
|
16510 |
+
|
16511 |
+
|
16512 |
+
|
16513 |
+
|
16514 |
+
|
16515 |
+
|
16516 |
+
|
16517 |
+
|
16518 |
+
|
16519 |
+
|
16520 |
+
|
16521 |
+
|
16522 |
+
|
16523 |
+
|
16524 |
+
|
16525 |
+
|
16526 |
+
|
16527 |
+
|
16528 |
+
|
16529 |
+
|
16530 |
+
|
16531 |
+
|
16532 |
+
|
16533 |
+
|
16534 |
+
|
16535 |
+
|
16536 |
+
|
16537 |
+
|
16538 |
+
|
16539 |
+
|
16540 |
+
|
16541 |
+
|
16542 |
+
|
16543 |
+
|
16544 |
+
|
16545 |
+
|
16546 |
+
|
16547 |
+
|
16548 |
+
|
16549 |
+
|
16550 |
+
|
16551 |
+
|
16552 |
+
|
16553 |
+
|
16554 |
+
|
16555 |
+
|
16556 |
+
|
16557 |
+
|
16558 |
+
|
16559 |
+
|
16560 |
+
|
16561 |
+
|
16562 |
+
|
16563 |
+
|
16564 |
+
|
16565 |
+
|
16566 |
+
|
16567 |
+
|
16568 |
+
|
16569 |
+
|
16570 |
+
|
16571 |
+
|
16572 |
+
|
16573 |
+
|
16574 |
+
|
16575 |
+
|
16576 |
+
|
16577 |
+
|
16578 |
+
|
16579 |
+
|
16580 |
+
|
16581 |
+
|
16582 |
+
|
16583 |
+
|
16584 |
+
|
16585 |
+
|
16586 |
+
|
16587 |
+
|
16588 |
+
|
16589 |
+
|
16590 |
+
|
16591 |
+
|
16592 |
+
|
16593 |
+
|
16594 |
+
|
16595 |
+
|
16596 |
+
|
16597 |
+
|
16598 |
+
|
16599 |
+
|
16600 |
+
|
16601 |
+
|
16602 |
+
|
16603 |
+
|
16604 |
+
|
16605 |
+
|
16606 |
+
|
16607 |
+
|
16608 |
+
|
16609 |
+
|
16610 |
+
|
16611 |
+
|
16612 |
+
|
16613 |
+
|
16614 |
+
|
16615 |
+
|
16616 |
+
|
16617 |
+
|
16618 |
+
|
16619 |
+
|
16620 |
+
|
16621 |
+
|
16622 |
+
|
16623 |
+
|
16624 |
+
|
16625 |
+
|
16626 |
+
|
16627 |
+
|
16628 |
+
|
16629 |
+
|
16630 |
+
|
16631 |
+
|
16632 |
+
|
16633 |
+
|
16634 |
+
|
16635 |
+
|
16636 |
+
|
16637 |
+
|
16638 |
+
|
16639 |
+
|
16640 |
+
|
16641 |
+
|
16642 |
+
|
16643 |
+
|
16644 |
+
|
16645 |
+
|
16646 |
+
|
16647 |
+
|
16648 |
+
|
16649 |
+
|
16650 |
+
|
16651 |
+
|
16652 |
+
|
16653 |
+
|
16654 |
+
|
16655 |
+
|
16656 |
+
|
16657 |
+
|
16658 |
+
|
16659 |
+
|
16660 |
+
|
16661 |
+
|
16662 |
+
|
16663 |
+
|
16664 |
+
|
16665 |
+
|
16666 |
+
|
16667 |
+
|
16668 |
+
|
16669 |
+
|
16670 |
+
|
16671 |
+
|
16672 |
+
|
16673 |
+
|
16674 |
+
|
16675 |
+
|
16676 |
+
|
16677 |
+
|
16678 |
+
|
16679 |
+
|
16680 |
+
|
16681 |
+
|
16682 |
+
|
16683 |
+
|
16684 |
+
|
16685 |
+
|
16686 |
+
|
16687 |
+
|
16688 |
+
|
16689 |
+
|
16690 |
+
|
16691 |
+
|
16692 |
+
|
16693 |
+
|
16694 |
+
|
16695 |
+
|
16696 |
+
|
16697 |
+
|
16698 |
+
|
16699 |
+
|
16700 |
+
|
16701 |
+
|
16702 |
+
|
16703 |
+
|
16704 |
+
|
16705 |
+
|
16706 |
+
|
16707 |
+
|
16708 |
+
|
16709 |
+
|
16710 |
+
|
16711 |
+
|
16712 |
+
|
16713 |
+
|
16714 |
+
|
16715 |
+
|
16716 |
+
|
16717 |
+
|
16718 |
+
|
16719 |
+
|
16720 |
+
|
16721 |
+
|
16722 |
+
|
16723 |
+
|
16724 |
+
|
16725 |
+
|
16726 |
+
|
16727 |
+
|
16728 |
+
|
16729 |
+
|
16730 |
+
|
16731 |
+
|
16732 |
+
|
16733 |
+
|
16734 |
+
|
16735 |
+
|
16736 |
+
|
16737 |
+
|
16738 |
+
|
16739 |
+
|
16740 |
+
|
16741 |
+
|
16742 |
+
|
16743 |
+
|
16744 |
+
|
16745 |
+
|
16746 |
+
|
16747 |
+
|
16748 |
+
|
16749 |
+
|
16750 |
+
|
16751 |
+
|
16752 |
+
|
16753 |
+
|
16754 |
+
|
16755 |
+
|
16756 |
+
|
16757 |
+
|
16758 |
+
|
16759 |
+
|
16760 |
+
|
16761 |
+
|
16762 |
+
|
16763 |
+
|
16764 |
+
|
16765 |
+
|
16766 |
+
|
16767 |
+
|
16768 |
+
|
16769 |
+
|
16770 |
+
|
16771 |
+
|
16772 |
+
|
16773 |
+
|
16774 |
+
|
16775 |
+
|
16776 |
+
|
16777 |
+
|
16778 |
+
|
16779 |
+
|
16780 |
+
|
16781 |
+
|
16782 |
+
|
16783 |
+
|
16784 |
+
|
16785 |
+
|
16786 |
+
|
16787 |
+
|
16788 |
+
|
16789 |
+
|
16790 |
+
|
16791 |
+
|
16792 |
+
|
16793 |
+
|
16794 |
+
|
16795 |
+
|
16796 |
+
|
16797 |
+
|
16798 |
+
|
16799 |
+
|
16800 |
+
|
16801 |
+
|
16802 |
+
|
16803 |
+
|
16804 |
+
|
16805 |
+
|
16806 |
+
|
16807 |
+
|
16808 |
+
|
16809 |
+
|
16810 |
+
|
16811 |
+
|
16812 |
+
|
16813 |
+
|
16814 |
+
|
16815 |
+
|
16816 |
+
|
16817 |
+
|
16818 |
+
|
16819 |
+
|
16820 |
+
|
16821 |
+
|
16822 |
+
|
16823 |
+
|
16824 |
+
|
16825 |
+
|
16826 |
+
|
16827 |
+
|
16828 |
+
|
16829 |
+
|
16830 |
+
|
16831 |
+
|
16832 |
+
|
16833 |
+
|
16834 |
+
|
16835 |
+
|
16836 |
+
|
16837 |
+
|
16838 |
+
|
16839 |
+
|
16840 |
+
|
16841 |
+
|
16842 |
+
|
16843 |
+
|
16844 |
+
|
16845 |
+
|
16846 |
+
|
16847 |
+
|
16848 |
+
|
16849 |
+
|
16850 |
+
|
16851 |
+
|
16852 |
+
|
16853 |
+
|
16854 |
+
|
16855 |
+
|
16856 |
+
|
16857 |
+
|
16858 |
+
|
16859 |
+
|
16860 |
+
|
16861 |
+
|
16862 |
+
|
16863 |
+
|
16864 |
+
|
16865 |
+
|
16866 |
+
|
16867 |
+
|
16868 |
+
|
16869 |
+
|
16870 |
+
|
16871 |
+
|
16872 |
+
|
16873 |
+
|
16874 |
+
|
16875 |
+
|
16876 |
+
|
16877 |
+
|
16878 |
+
|
16879 |
+
|
16880 |
+
|
16881 |
+
|
16882 |
+
|
16883 |
+
|
16884 |
+
|
16885 |
+
|
16886 |
+
|
16887 |
+
|
16888 |
+
|
16889 |
+
|
16890 |
+
|
16891 |
+
|
16892 |
+
|
16893 |
+
|
16894 |
+
|
16895 |
+
|
16896 |
+
|
16897 |
+
|
16898 |
+
|
16899 |
+
|
16900 |
+
|
16901 |
+
|
16902 |
+
|
16903 |
+
|
16904 |
+
|
16905 |
+
|
16906 |
+
|
16907 |
+
|
16908 |
+
|
16909 |
+
|
16910 |
+
|
16911 |
+
|
16912 |
+
|
16913 |
+
Step... (24000/50000 | Loss: 1.6508632898330688, Acc: 0.6671841740608215): 50%|█████████████▌ | 25000/50000 [9:54:56<11:05:00, 1.60s/it]
|
16914 |
+
Evaluating ...: 5%|████▍ | 6/130 [00:00<00:07, 15.90it/s]
|
16915 |
+
Step... (24500 | Loss: 1.7519614696502686, Learning Rate: 0.0003090909158345312)
|
16916 |
+
|
16917 |
+
|
16918 |
+
|
16919 |
+
|
16920 |
+
|
16921 |
+
|
16922 |
+
|
16923 |
+
|
16924 |
+
|
16925 |
+
|
16926 |
+
|
16927 |
+
|
16928 |
+
[12:03:19] - INFO - __main__ - Saving checkpoint at 25000 steps█████████████████████████████████████████████████████| 130/130 [00:21<00:00, 4.60it/s]
|
16929 |
+
All Flax model weights were used when initializing RobertaForMaskedLM.
|
16930 |
+
Some weights of RobertaForMaskedLM were not initialized from the Flax model and are newly initialized: ['lm_head.decoder.weight', 'roberta.embeddings.position_ids', 'lm_head.decoder.bias']
|
16931 |
+
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
|
16932 |
+
|
16933 |
+
|
16934 |
+
|
16935 |
+
|
16936 |
+
|
16937 |
+
|
16938 |
+
|
16939 |
+
|
16940 |
+
|
16941 |
+
|
16942 |
+
|
16943 |
+
|
16944 |
+
|
16945 |
+
|
16946 |
+
|
16947 |
+
|
16948 |
+
|
16949 |
+
|
16950 |
+
|
16951 |
+
|
16952 |
+
|
16953 |
+
|
16954 |
+
|
16955 |
+
|
16956 |
+
|
16957 |
+
|
16958 |
+
|
16959 |
+
|
16960 |
+
|
16961 |
+
|
16962 |
+
|
16963 |
+
|
16964 |
+
|
16965 |
+
|
16966 |
+
|
16967 |
+
|
16968 |
+
|
16969 |
+
|
16970 |
+
|
16971 |
+
|
16972 |
+
|
16973 |
+
|
16974 |
+
|
16975 |
+
|
16976 |
+
|
16977 |
+
|
16978 |
+
|
16979 |
+
|
16980 |
+
|
16981 |
+
|
16982 |
+
|
16983 |
+
|
16984 |
+
|
16985 |
+
|
16986 |
+
|
16987 |
+
|
16988 |
+
|
16989 |
+
|
16990 |
+
|
16991 |
+
|
16992 |
+
|
16993 |
+
|
16994 |
+
|
16995 |
+
|
16996 |
+
|
16997 |
+
|
16998 |
+
|
16999 |
+
|
17000 |
+
|
17001 |
+
|
17002 |
+
|
17003 |
+
|
17004 |
+
|
17005 |
+
|
17006 |
+
|
17007 |
+
|
17008 |
+
|
17009 |
+
|
17010 |
+
|
17011 |
+
|
17012 |
+
|
17013 |
+
|
17014 |
+
|
17015 |
+
|
17016 |
+
|
17017 |
+
|
17018 |
+
|
17019 |
+
|
17020 |
+
|
17021 |
+
|
17022 |
+
|
17023 |
+
|
17024 |
+
|
17025 |
+
|
17026 |
+
|
17027 |
+
|
17028 |
+
|
17029 |
+
|
17030 |
+
|
17031 |
+
|
17032 |
+
|
17033 |
+
|
17034 |
+
|
17035 |
+
|
17036 |
+
|
17037 |
+
|
17038 |
+
|
17039 |
+
|
17040 |
+
|
17041 |
+
|
17042 |
+
|
17043 |
+
|
17044 |
+
|
17045 |
+
|
17046 |
+
|
17047 |
+
|
17048 |
+
|
17049 |
+
|
17050 |
+
|
17051 |
+
|
17052 |
+
|
17053 |
+
|
17054 |
+
|
17055 |
+
|
17056 |
+
|
17057 |
+
|
17058 |
+
|
17059 |
+
|
17060 |
+
|
17061 |
+
|
17062 |
+
|
17063 |
+
|
17064 |
+
|
17065 |
+
|
17066 |
+
|
17067 |
+
|
17068 |
+
|
17069 |
+
|
17070 |
+
|
17071 |
+
|
17072 |
+
|
17073 |
+
|
17074 |
+
|
17075 |
+
|
17076 |
+
|
17077 |
+
|
17078 |
+
|
17079 |
+
|
17080 |
+
|
17081 |
+
|
17082 |
+
|
17083 |
+
|
17084 |
+
|
17085 |
+
|
17086 |
+
|
17087 |
+
|
17088 |
+
|
17089 |
+
|
17090 |
+
|
17091 |
+
|
17092 |
+
|
17093 |
+
|
17094 |
+
|
17095 |
+
|
17096 |
+
|
17097 |
+
|
17098 |
+
|
17099 |
+
|
17100 |
+
|
17101 |
+
|
17102 |
+
|
17103 |
+
|
17104 |
+
|
17105 |
+
|
17106 |
+
|
17107 |
+
|
17108 |
+
|
17109 |
+
|
17110 |
+
|
17111 |
+
|
17112 |
+
|
17113 |
+
|
17114 |
+
|
17115 |
+
|
17116 |
+
|
17117 |
+
|
17118 |
+
|
17119 |
+
|
17120 |
+
|
17121 |
+
|
17122 |
+
|
17123 |
+
|
17124 |
+
|
17125 |
+
|
17126 |
+
|
17127 |
+
|
17128 |
+
|
17129 |
+
|
17130 |
+
|
17131 |
+
|
17132 |
+
|
17133 |
+
|
17134 |
+
|
17135 |
+
|
17136 |
+
|
17137 |
+
|
17138 |
+
|
17139 |
+
|
17140 |
+
|
17141 |
+
|
17142 |
+
|
17143 |
+
|
17144 |
+
|
17145 |
+
|
17146 |
+
|
17147 |
+
|
17148 |
+
|
17149 |
+
|
17150 |
+
|
17151 |
+
|
17152 |
+
|
17153 |
+
|
17154 |
+
|
17155 |
+
|
17156 |
+
|
17157 |
+
|
17158 |
+
|
17159 |
+
|
17160 |
+
|
17161 |
+
|
17162 |
+
|
17163 |
+
|
17164 |
+
|
17165 |
+
|
17166 |
+
|
17167 |
+
|
17168 |
+
|
17169 |
+
|
17170 |
+
|
17171 |
+
|
17172 |
+
|
17173 |
+
|
17174 |
+
|
17175 |
+
|
17176 |
+
|
17177 |
+
|
17178 |
+
|
17179 |
+
|
17180 |
+
|
17181 |
+
|
17182 |
+
|
17183 |
+
|
17184 |
+
|
17185 |
+
|
17186 |
+
|
17187 |
+
|
17188 |
+
|
17189 |
+
|
17190 |
+
|
17191 |
+
|
17192 |
+
|
17193 |
+
|
17194 |
+
|
17195 |
+
|
17196 |
+
|
17197 |
+
|
17198 |
+
|
17199 |
+
|
17200 |
+
|
17201 |
+
|
17202 |
+
|
17203 |
+
|
17204 |
+
|
17205 |
+
|
17206 |
+
|
17207 |
+
|
17208 |
+
|
17209 |
+
|
17210 |
+
|
17211 |
+
|
17212 |
+
|
17213 |
+
|
17214 |
+
|
17215 |
+
|
17216 |
+
|
17217 |
+
|
17218 |
+
|
17219 |
+
|
17220 |
+
|
17221 |
+
|
17222 |
+
|
17223 |
+
|
17224 |
+
|
17225 |
+
|
17226 |
+
|
17227 |
+
|
17228 |
+
|
17229 |
+
|
17230 |
+
|
17231 |
+
|
17232 |
+
|
17233 |
+
|
17234 |
+
|
17235 |
+
|
17236 |
+
|
17237 |
+
|
17238 |
+
|
17239 |
+
|
17240 |
+
|
17241 |
+
|
17242 |
+
|
17243 |
+
|
17244 |
+
|
17245 |
+
|
17246 |
+
|
17247 |
+
|
17248 |
+
|
17249 |
+
|
17250 |
+
|
17251 |
+
|
17252 |
+
|
17253 |
+
|
17254 |
+
|
17255 |
+
|
17256 |
+
|
17257 |
+
|
17258 |
+
|
17259 |
+
|
17260 |
+
|
17261 |
+
|
17262 |
+
|
17263 |
+
|
17264 |
+
|
17265 |
+
|
17266 |
+
|
17267 |
+
|
17268 |
+
|
17269 |
+
|
17270 |
+
|
17271 |
+
|
17272 |
+
|
17273 |
+
|
17274 |
+
|
17275 |
+
|
17276 |
+
|
17277 |
+
|
17278 |
+
|
17279 |
+
|
17280 |
+
|
17281 |
+
|
17282 |
+
|
17283 |
+
|
17284 |
+
|
17285 |
+
|
17286 |
+
|
17287 |
+
|
17288 |
+
|
17289 |
+
|
17290 |
+
|
17291 |
+
|
17292 |
+
|
17293 |
+
|
17294 |
+
|
17295 |
+
|
17296 |
+
|
17297 |
+
|
17298 |
+
|
17299 |
+
|
17300 |
+
|
17301 |
+
|
17302 |
+
|
17303 |
+
|
17304 |
+
|
17305 |
+
|
17306 |
+
|
17307 |
+
|
17308 |
+
|
17309 |
+
|
17310 |
+
|
17311 |
+
|
17312 |
+
|
17313 |
+
|
17314 |
+
|
17315 |
+
|
17316 |
+
|
17317 |
+
|
17318 |
+
|
17319 |
+
|
17320 |
+
|
17321 |
+
|
17322 |
+
|
17323 |
+
|
17324 |
+
|
17325 |
+
|
17326 |
+
|
17327 |
+
|
17328 |
+
|
17329 |
+
|
17330 |
+
|
17331 |
+
|
17332 |
+
|
17333 |
+
|
17334 |
+
|
17335 |
+
|
17336 |
+
|
17337 |
+
|
17338 |
+
|
17339 |
+
|
17340 |
+
|
17341 |
+
|
17342 |
+
|
17343 |
+
|
17344 |
+
|
17345 |
+
|
17346 |
+
|
17347 |
+
|
17348 |
+
|
17349 |
+
|
17350 |
+
|
17351 |
+
|
17352 |
+
|
17353 |
+
|
17354 |
+
|
17355 |
+
|
17356 |
+
|
17357 |
+
|
17358 |
+
|
17359 |
+
|
17360 |
+
|
17361 |
+
|
17362 |
+
|
17363 |
+
|
17364 |
+
|
17365 |
+
|
17366 |
+
|
17367 |
+
|
17368 |
+
|
17369 |
+
|
17370 |
+
|
17371 |
+
|
17372 |
+
|
17373 |
+
|
17374 |
+
|
17375 |
+
|
17376 |
+
|
17377 |
+
|
17378 |
+
|
17379 |
+
|
17380 |
+
|
17381 |
+
|
17382 |
+
|
17383 |
+
|
17384 |
+
|
17385 |
+
|
17386 |
+
|
17387 |
+
|
17388 |
+
|
17389 |
+
|
17390 |
+
|
17391 |
+
|
17392 |
+
|
17393 |
+
|
17394 |
+
|
17395 |
+
|
17396 |
+
|
17397 |
+
|
17398 |
+
|
17399 |
+
|
17400 |
+
|
17401 |
+
|
17402 |
+
|
17403 |
+
|
17404 |
+
|
17405 |
+
|
17406 |
+
|
17407 |
+
|
17408 |
+
|
17409 |
+
|
17410 |
+
|
17411 |
+
|
17412 |
+
|
17413 |
+
|
17414 |
+
|
17415 |
+
|
17416 |
+
|
17417 |
+
|
17418 |
+
|
17419 |
+
|
17420 |
+
|
17421 |
+
|
17422 |
+
|
17423 |
+
|
17424 |
+
|
17425 |
+
|
17426 |
+
|
17427 |
+
|
17428 |
+
|
17429 |
+
|
17430 |
+
|
17431 |
+
|
17432 |
+
|
17433 |
+
|
17434 |
+
|
17435 |
+
|
17436 |
+
|
17437 |
+
|
17438 |
+
|
17439 |
+
|
17440 |
+
|
17441 |
+
|
17442 |
+
|
17443 |
+
|
17444 |
+
|
17445 |
+
|
17446 |
+
|
17447 |
+
|
17448 |
+
|
17449 |
+
|
17450 |
+
|
17451 |
+
|
17452 |
+
|
17453 |
+
|
17454 |
+
|
17455 |
+
|
17456 |
+
|
17457 |
+
|
17458 |
+
|
17459 |
+
|
17460 |
+
|
17461 |
+
|
17462 |
+
|
17463 |
+
|
17464 |
+
|
17465 |
+
|
17466 |
+
|
17467 |
+
|
17468 |
+
|
17469 |
+
|
17470 |
+
|
17471 |
+
|
17472 |
+
|
17473 |
+
|
17474 |
+
|
17475 |
+
|
17476 |
+
|
17477 |
+
|
17478 |
+
|
17479 |
+
|
17480 |
+
|
17481 |
+
|
17482 |
+
|
17483 |
+
|
17484 |
+
|
17485 |
+
|
17486 |
+
|
17487 |
+
|
17488 |
+
|
17489 |
+
|
17490 |
+
|
17491 |
+
|
17492 |
+
|
17493 |
+
|
17494 |
+
|
17495 |
+
|
17496 |
+
|
17497 |
+
|
17498 |
+
|
17499 |
+
|
17500 |
+
|
17501 |
+
|
17502 |
+
|
17503 |
+
|
17504 |
+
|
17505 |
+
|
17506 |
+
|
17507 |
+
|
17508 |
+
|
17509 |
+
|
17510 |
+
|
17511 |
+
|
17512 |
+
|
17513 |
+
|
17514 |
+
|
17515 |
+
|
17516 |
+
|
17517 |
+
|
17518 |
+
|
17519 |
+
|
17520 |
+
|
17521 |
+
|
17522 |
+
|
17523 |
+
|
17524 |
+
|
17525 |
+
|
17526 |
+
|
17527 |
+
|
17528 |
+
|
17529 |
+
|
17530 |
+
|
17531 |
+
|
17532 |
+
|
17533 |
+
|
17534 |
+
|
17535 |
+
|
17536 |
+
|
17537 |
+
|
17538 |
+
|
17539 |
+
|
17540 |
+
|
17541 |
+
|
17542 |
+
|
17543 |
+
|
17544 |
+
|
17545 |
+
|
17546 |
+
|
17547 |
+
|
17548 |
+
|
17549 |
+
|
17550 |
+
|
17551 |
+
|
17552 |
+
|
17553 |
+
|
17554 |
+
|
17555 |
+
|
17556 |
+
|
17557 |
+
|
17558 |
+
|
17559 |
+
|
17560 |
+
|
17561 |
+
|
17562 |
+
|
17563 |
+
|
17564 |
+
|
17565 |
+
|
17566 |
+
|
17567 |
+
|
17568 |
+
|
17569 |
+
|
17570 |
+
|
17571 |
+
|
17572 |
+
|
17573 |
+
|
17574 |
+
|
17575 |
+
|
17576 |
+
|
17577 |
+
|
17578 |
+
|
17579 |
+
|
17580 |
+
|
17581 |
+
|
17582 |
+
|
17583 |
+
|
17584 |
+
|
17585 |
+
|
17586 |
+
|
17587 |
+
|
17588 |
+
|
17589 |
+
|
17590 |
+
|
17591 |
+
|
17592 |
+
|
17593 |
+
|
17594 |
+
|
17595 |
+
|
17596 |
+
|
17597 |
+
|
17598 |
+
|
17599 |
+
|
17600 |
+
|
17601 |
+
|
17602 |
+
|
17603 |
+
|
17604 |
+
|
17605 |
+
|
17606 |
+
|
17607 |
+
|
17608 |
+
|
17609 |
+
|
17610 |
+
|
17611 |
+
|
17612 |
+
|
17613 |
+
Step... (25000/50000 | Loss: 1.6436606645584106, Acc: 0.668701171875): 52%|████████████████ | 26000/50000 [10:20:04<8:52:22, 1.33s/it]
|
17614 |
+
Step... (25500 | Loss: 1.6520822048187256, Learning Rate: 0.0002969697234220803)
|
17615 |
+
Step... (26000 | Loss: 1.7167686223983765, Learning Rate: 0.0002909091126639396)
|
17616 |
+
|
17617 |
+
|
17618 |
+
|
17619 |
+
|
17620 |
+
|
17621 |
+
|
17622 |
+
|
17623 |
+
|
17624 |
+
|
17625 |
+
|
17626 |
+
|
17627 |
+
[12:28:28] - INFO - __main__ - Saving checkpoint at 26000 steps█████████████████████████████████████████████████████| 130/130 [00:21<00:00, 4.60it/s]
|
17628 |
+
All Flax model weights were used when initializing RobertaForMaskedLM.
|
17629 |
+
Some weights of RobertaForMaskedLM were not initialized from the Flax model and are newly initialized: ['lm_head.decoder.weight', 'roberta.embeddings.position_ids', 'lm_head.decoder.bias']
|
17630 |
+
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
|
17631 |
+
|
17632 |
+
|
17633 |
+
|
17634 |
+
|
17635 |
+
|
17636 |
+
|
17637 |
+
|
17638 |
+
|
17639 |
+
|
17640 |
+
|
17641 |
+
|
17642 |
+
|
17643 |
+
|
17644 |
+
|
17645 |
+
|
17646 |
+
|
17647 |
+
|
17648 |
+
|
17649 |
+
|
17650 |
+
|
17651 |
+
|
17652 |
+
|
17653 |
+
|
17654 |
+
|
17655 |
+
|
17656 |
+
|
17657 |
+
|
17658 |
+
|
17659 |
+
|
17660 |
+
|
17661 |
+
|
17662 |
+
|
17663 |
+
|
17664 |
+
|
17665 |
+
|
17666 |
+
|
17667 |
+
|
17668 |
+
|
17669 |
+
|
17670 |
+
|
17671 |
+
|
17672 |
+
|
17673 |
+
|
17674 |
+
|
17675 |
+
|
17676 |
+
|
17677 |
+
|
17678 |
+
|
17679 |
+
|
17680 |
+
|
17681 |
+
|
17682 |
+
|
17683 |
+
|
17684 |
+
|
17685 |
+
|
17686 |
+
|
17687 |
+
|
17688 |
+
|
17689 |
+
|
17690 |
+
|
17691 |
+
|
17692 |
+
|
17693 |
+
|
17694 |
+
|
17695 |
+
|
17696 |
+
|
17697 |
+
|
17698 |
+
|
17699 |
+
|
17700 |
+
|
17701 |
+
|
17702 |
+
|
17703 |
+
|
17704 |
+
|
17705 |
+
|
17706 |
+
|
17707 |
+
|
17708 |
+
|
17709 |
+
|
17710 |
+
|
17711 |
+
|
17712 |
+
|
17713 |
+
|
17714 |
+
|
17715 |
+
|
17716 |
+
|
17717 |
+
|
17718 |
+
|
17719 |
+
|
17720 |
+
|
17721 |
+
|
17722 |
+
|
17723 |
+
|
17724 |
+
|
17725 |
+
|
17726 |
+
|
17727 |
+
|
17728 |
+
|
17729 |
+
|
17730 |
+
|
17731 |
+
|
17732 |
+
|
17733 |
+
|
17734 |
+
|
17735 |
+
|
17736 |
+
|
17737 |
+
|
17738 |
+
|
17739 |
+
|
17740 |
+
|
17741 |
+
|
17742 |
+
|
17743 |
+
|
17744 |
+
|
17745 |
+
|
17746 |
+
|
17747 |
+
|
17748 |
+
|
17749 |
+
|
17750 |
+
|
17751 |
+
|
17752 |
+
|
17753 |
+
|
17754 |
+
|
17755 |
+
|
17756 |
+
|
17757 |
+
|
17758 |
+
|
17759 |
+
|
17760 |
+
|
17761 |
+
|
17762 |
+
|
17763 |
+
|
17764 |
+
|
17765 |
+
|
17766 |
+
|
17767 |
+
|
17768 |
+
|
17769 |
+
|
17770 |
+
|
17771 |
+
|
17772 |
+
|
17773 |
+
|
17774 |
+
|
17775 |
+
|
17776 |
+
|
17777 |
+
|
17778 |
+
|
17779 |
+
|
17780 |
+
|
17781 |
+
|
17782 |
+
|
17783 |
+
|
17784 |
+
|
17785 |
+
|
17786 |
+
|
17787 |
+
|
17788 |
+
|
17789 |
+
|
17790 |
+
|
17791 |
+
|
17792 |
+
|
17793 |
+
|
17794 |
+
|
17795 |
+
|
17796 |
+
|
17797 |
+
|
17798 |
+
|
17799 |
+
|
17800 |
+
|
17801 |
+
|
17802 |
+
|
17803 |
+
|
17804 |
+
|
17805 |
+
|
17806 |
+
|
17807 |
+
|
17808 |
+
|
17809 |
+
|
17810 |
+
|
17811 |
+
|
17812 |
+
|
17813 |
+
|
17814 |
+
|
17815 |
+
|
17816 |
+
|
17817 |
+
|
17818 |
+
|
17819 |
+
|
17820 |
+
|
17821 |
+
|
17822 |
+
|
17823 |
+
|
17824 |
+
|
17825 |
+
|
17826 |
+
|
17827 |
+
|
17828 |
+
|
17829 |
+
|
17830 |
+
|
17831 |
+
|
17832 |
+
|
17833 |
+
|
17834 |
+
|
17835 |
+
|
17836 |
+
|
17837 |
+
|
17838 |
+
|
17839 |
+
|
17840 |
+
|
17841 |
+
|
17842 |
+
|
17843 |
+
|
17844 |
+
|
17845 |
+
|
17846 |
+
|
17847 |
+
|
17848 |
+
|
17849 |
+
|
17850 |
+
|
17851 |
+
|
17852 |
+
|
17853 |
+
|
17854 |
+
|
17855 |
+
|
17856 |
+
|
17857 |
+
|
17858 |
+
|
17859 |
+
|
17860 |
+
|
17861 |
+
|
17862 |
+
|
17863 |
+
|
17864 |
+
|
17865 |
+
|
17866 |
+
|
17867 |
+
|
17868 |
+
|
17869 |
+
|
17870 |
+
|
17871 |
+
|
17872 |
+
|
17873 |
+
|
17874 |
+
|
17875 |
+
|
17876 |
+
|
17877 |
+
|
17878 |
+
|
17879 |
+
|
17880 |
+
|
17881 |
+
|
17882 |
+
|
17883 |
+
|
17884 |
+
|
17885 |
+
|
17886 |
+
|
17887 |
+
|
17888 |
+
|
17889 |
+
|
17890 |
+
|
17891 |
+
|
17892 |
+
|
17893 |
+
|
17894 |
+
|
17895 |
+
|
17896 |
+
|
17897 |
+
|
17898 |
+
|
17899 |
+
|
17900 |
+
|
17901 |
+
|
17902 |
+
|
17903 |
+
|
17904 |
+
|
17905 |
+
|
17906 |
+
|
17907 |
+
|
17908 |
+
|
17909 |
+
|
17910 |
+
|
17911 |
+
|
17912 |
+
|
17913 |
+
|
17914 |
+
|
17915 |
+
|
17916 |
+
|
17917 |
+
|
17918 |
+
|
17919 |
+
|
17920 |
+
|
17921 |
+
|
17922 |
+
|
17923 |
+
|
17924 |
+
|
17925 |
+
|
17926 |
+
|
17927 |
+
|
17928 |
+
|
17929 |
+
|
17930 |
+
|
17931 |
+
|
17932 |
+
|
17933 |
+
|
17934 |
+
|
17935 |
+
|
17936 |
+
|
17937 |
+
|
17938 |
+
|
17939 |
+
|
17940 |
+
|
17941 |
+
|
17942 |
+
|
17943 |
+
|
17944 |
+
|
17945 |
+
|
17946 |
+
|
17947 |
+
|
17948 |
+
|
17949 |
+
|
17950 |
+
|
17951 |
+
|
17952 |
+
|
17953 |
+
|
17954 |
+
|
17955 |
+
|
17956 |
+
|
17957 |
+
|
17958 |
+
|
17959 |
+
|
17960 |
+
|
17961 |
+
|
17962 |
+
|
17963 |
+
|
17964 |
+
|
17965 |
+
|
17966 |
+
|
17967 |
+
|
17968 |
+
|
17969 |
+
|
17970 |
+
|
17971 |
+
|
17972 |
+
|
17973 |
+
|
17974 |
+
|
17975 |
+
|
17976 |
+
|
17977 |
+
|
17978 |
+
|
17979 |
+
|
17980 |
+
|
17981 |
+
|
17982 |
+
|
17983 |
+
|
17984 |
+
|
17985 |
+
|
17986 |
+
|
17987 |
+
|
17988 |
+
|
17989 |
+
|
17990 |
+
|
17991 |
+
|
17992 |
+
|
17993 |
+
|
17994 |
+
|
17995 |
+
|
17996 |
+
|
17997 |
+
|
17998 |
+
|
17999 |
+
|
18000 |
+
|
18001 |
+
|
18002 |
+
|
18003 |
+
|
18004 |
+
|
18005 |
+
|
18006 |
+
|
18007 |
+
|
18008 |
+
|
18009 |
+
|
18010 |
+
|
18011 |
+
|
18012 |
+
|
18013 |
+
|
18014 |
+
|
18015 |
+
|
18016 |
+
|
18017 |
+
|
18018 |
+
|
18019 |
+
|
18020 |
+
|
18021 |
+
|
18022 |
+
|
18023 |
+
|
18024 |
+
|
18025 |
+
|
18026 |
+
|
18027 |
+
|
18028 |
+
|
18029 |
+
|
18030 |
+
|
18031 |
+
|
18032 |
+
|
18033 |
+
|
18034 |
+
|
18035 |
+
|
18036 |
+
|
18037 |
+
|
18038 |
+
|
18039 |
+
|
18040 |
+
|
18041 |
+
|
18042 |
+
|
18043 |
+
|
18044 |
+
|
18045 |
+
|
18046 |
+
|
18047 |
+
|
18048 |
+
|
18049 |
+
|
18050 |
+
|
18051 |
+
|
18052 |
+
|
18053 |
+
|
18054 |
|
18055 |
|
18056 |
|
wandb/run-20210726_001233-17u6inbn/files/wandb-summary.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"global_step":
|
|
|
1 |
+
{"global_step": 26500, "_timestamp": 1627303266.586647, "train_time": 1372972.75, "train_learning_rate": 0.0002848485019057989, "_step": 52841, "train_loss": 1.730733036994934, "eval_accuracy": 0.6691190600395203, "eval_loss": 1.6362030506134033}
|
wandb/run-20210726_001233-17u6inbn/logs/debug-internal.log
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:45d92a70b4478861c91b2901d46e09c9d061fd6a02da2c0ff8ca5335c33cfde8
|
3 |
+
size 20913336
|
wandb/run-20210726_001233-17u6inbn/run-17u6inbn.wandb
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1343d0ef55818f25efb5ed042e2cf55bb6c6f3440aba38c3b2f266129989e652
|
3 |
+
size 10444427
|