versae commited on
Commit
a015097
·
1 Parent(s): de8489d

Step... (26000/50000 | Loss: 1.6362030506134033, Acc: 0.6691190600395203): 53%|█████████████▊ | 26656/50000 [10:36:57<11:06:38, 1.71s/it]

Browse files
Files changed (25) hide show
  1. flax_model.msgpack +1 -1
  2. outputs/checkpoints/checkpoint-20000/training_state.json +0 -1
  3. outputs/checkpoints/checkpoint-21000/training_state.json +0 -1
  4. outputs/checkpoints/{checkpoint-20000 → checkpoint-25000}/config.json +0 -0
  5. outputs/checkpoints/{checkpoint-20000 → checkpoint-25000}/data_collator.joblib +0 -0
  6. outputs/checkpoints/{checkpoint-20000 → checkpoint-25000}/flax_model.msgpack +1 -1
  7. outputs/checkpoints/{checkpoint-20000 → checkpoint-25000}/optimizer_state.msgpack +1 -1
  8. outputs/checkpoints/{checkpoint-20000 → checkpoint-25000}/training_args.joblib +0 -0
  9. outputs/checkpoints/checkpoint-25000/training_state.json +1 -0
  10. outputs/checkpoints/{checkpoint-21000 → checkpoint-26000}/config.json +0 -0
  11. outputs/checkpoints/{checkpoint-21000 → checkpoint-26000}/data_collator.joblib +0 -0
  12. outputs/checkpoints/{checkpoint-21000 → checkpoint-26000}/flax_model.msgpack +1 -1
  13. outputs/checkpoints/{checkpoint-21000 → checkpoint-26000}/optimizer_state.msgpack +1 -1
  14. outputs/checkpoints/{checkpoint-21000 → checkpoint-26000}/training_args.joblib +0 -0
  15. outputs/checkpoints/checkpoint-26000/training_state.json +1 -0
  16. outputs/events.out.tfevents.1627258355.tablespoon.3000110.3.v2 +2 -2
  17. outputs/flax_model.msgpack +1 -1
  18. outputs/optimizer_state.msgpack +1 -1
  19. outputs/training_state.json +1 -1
  20. pytorch_model.bin +1 -1
  21. run_stream.512.log +0 -0
  22. wandb/run-20210726_001233-17u6inbn/files/output.log +1704 -0
  23. wandb/run-20210726_001233-17u6inbn/files/wandb-summary.json +1 -1
  24. wandb/run-20210726_001233-17u6inbn/logs/debug-internal.log +2 -2
  25. wandb/run-20210726_001233-17u6inbn/run-17u6inbn.wandb +2 -2
flax_model.msgpack CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b22d22612dd38ad92ffdda4b0cf432e201d6c90dd5386d04a2cdf4d19cdfd1ed
3
  size 249750019
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ceb7903825a5802361e2abfa48cc22549619333ade477fa0502eca94fff7518
3
  size 249750019
outputs/checkpoints/checkpoint-20000/training_state.json DELETED
@@ -1 +0,0 @@
1
- {"step": 20001}
 
 
outputs/checkpoints/checkpoint-21000/training_state.json DELETED
@@ -1 +0,0 @@
1
- {"step": 21001}
 
 
outputs/checkpoints/{checkpoint-20000 → checkpoint-25000}/config.json RENAMED
File without changes
outputs/checkpoints/{checkpoint-20000 → checkpoint-25000}/data_collator.joblib RENAMED
File without changes
outputs/checkpoints/{checkpoint-20000 → checkpoint-25000}/flax_model.msgpack RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a0bf6ec68f574b4af0a99a55d1d45aa7b0e1e57289109e4d7f5c9e5d83816424
3
  size 249750019
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f01f7bd9dc37008fe586042d0858630facb5e28a5aabe227d941a460b9dac62
3
  size 249750019
outputs/checkpoints/{checkpoint-20000 → checkpoint-25000}/optimizer_state.msgpack RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cab56f5cbba1f1be7c19b45ad56f5c58ced35d8e3f342687638a6a00d3e1d292
3
  size 499500278
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd4e308d24cbf75134aac7314daf08bd6739fe89d8e0c1dc9de2428e0a4a588a
3
  size 499500278
outputs/checkpoints/{checkpoint-20000 → checkpoint-25000}/training_args.joblib RENAMED
File without changes
outputs/checkpoints/checkpoint-25000/training_state.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"step": 25001}
outputs/checkpoints/{checkpoint-21000 → checkpoint-26000}/config.json RENAMED
File without changes
outputs/checkpoints/{checkpoint-21000 → checkpoint-26000}/data_collator.joblib RENAMED
File without changes
outputs/checkpoints/{checkpoint-21000 → checkpoint-26000}/flax_model.msgpack RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7ba1daf7b1dad5bf7c386bc7b53d5537a8f26b3cfee5b0fc009a750ad077eab0
3
  size 249750019
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ceb7903825a5802361e2abfa48cc22549619333ade477fa0502eca94fff7518
3
  size 249750019
outputs/checkpoints/{checkpoint-21000 → checkpoint-26000}/optimizer_state.msgpack RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cd862c6893d8672a836d674b5ef9d3eaab357c385ad5b064b7202eccc581ff05
3
  size 499500278
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22861588524628232613a207d4ef751ae2280267840c912f9817827bd0747a14
3
  size 499500278
outputs/checkpoints/{checkpoint-21000 → checkpoint-26000}/training_args.joblib RENAMED
File without changes
outputs/checkpoints/checkpoint-26000/training_state.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"step": 26001}
outputs/events.out.tfevents.1627258355.tablespoon.3000110.3.v2 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:187bfd40e3dd6f12ab8cd6df2018b0fef55ab1ab89a973e1cc1b5427620d8135
3
- size 3549865
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c07c9e738eed842b50ab4e82610a59b50a7f87c6b9aeeff615946beccf6da7a2
3
+ size 3922999
outputs/flax_model.msgpack CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b22d22612dd38ad92ffdda4b0cf432e201d6c90dd5386d04a2cdf4d19cdfd1ed
3
  size 249750019
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ceb7903825a5802361e2abfa48cc22549619333ade477fa0502eca94fff7518
3
  size 249750019
outputs/optimizer_state.msgpack CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bcac7bac463ddd6530546523b0141118f658d528e0d7ec682da2661fe2a0f7df
3
  size 499500278
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22861588524628232613a207d4ef751ae2280267840c912f9817827bd0747a14
3
  size 499500278
outputs/training_state.json CHANGED
@@ -1 +1 @@
1
- {"step": 24001}
 
1
+ {"step": 26001}
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d50ca6bc265a7b18cee3972966e847d1c5891e5fec62a6e912bbbe885e2e82da
3
  size 498858859
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17b4d4ba780e464164915cfe3081ea36a32c7f31487d8cc141e99fe4a4e44171
3
  size 498858859
run_stream.512.log CHANGED
The diff for this file is too large to render. See raw diff
 
wandb/run-20210726_001233-17u6inbn/files/output.log CHANGED
@@ -16347,6 +16347,1710 @@ You should probably TRAIN this model on a down-stream task to be able to use it
16347
 
16348
 
16349
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16350
 
16351
 
16352
 
 
16347
 
16348
 
16349
 
16350
+
16351
+
16352
+
16353
+
16354
+
16355
+
16356
+
16357
+
16358
+
16359
+
16360
+
16361
+
16362
+
16363
+
16364
+
16365
+
16366
+
16367
+
16368
+
16369
+
16370
+
16371
+
16372
+
16373
+
16374
+
16375
+
16376
+
16377
+
16378
+
16379
+
16380
+
16381
+
16382
+
16383
+
16384
+
16385
+
16386
+
16387
+
16388
+
16389
+
16390
+
16391
+
16392
+
16393
+
16394
+
16395
+
16396
+
16397
+
16398
+
16399
+
16400
+
16401
+
16402
+
16403
+
16404
+
16405
+
16406
+
16407
+
16408
+
16409
+
16410
+
16411
+
16412
+
16413
+
16414
+
16415
+
16416
+
16417
+
16418
+
16419
+
16420
+
16421
+
16422
+
16423
+
16424
+
16425
+
16426
+
16427
+
16428
+
16429
+
16430
+
16431
+
16432
+
16433
+
16434
+
16435
+
16436
+
16437
+
16438
+
16439
+
16440
+
16441
+
16442
+
16443
+
16444
+
16445
+
16446
+
16447
+
16448
+
16449
+
16450
+
16451
+
16452
+
16453
+
16454
+
16455
+
16456
+
16457
+
16458
+
16459
+
16460
+
16461
+
16462
+
16463
+
16464
+
16465
+
16466
+
16467
+
16468
+
16469
+
16470
+
16471
+
16472
+
16473
+
16474
+
16475
+
16476
+
16477
+
16478
+
16479
+
16480
+
16481
+
16482
+
16483
+
16484
+
16485
+
16486
+
16487
+
16488
+
16489
+
16490
+
16491
+
16492
+
16493
+
16494
+
16495
+
16496
+
16497
+
16498
+
16499
+
16500
+
16501
+
16502
+
16503
+
16504
+
16505
+
16506
+
16507
+
16508
+
16509
+
16510
+
16511
+
16512
+
16513
+
16514
+
16515
+
16516
+
16517
+
16518
+
16519
+
16520
+
16521
+
16522
+
16523
+
16524
+
16525
+
16526
+
16527
+
16528
+
16529
+
16530
+
16531
+
16532
+
16533
+
16534
+
16535
+
16536
+
16537
+
16538
+
16539
+
16540
+
16541
+
16542
+
16543
+
16544
+
16545
+
16546
+
16547
+
16548
+
16549
+
16550
+
16551
+
16552
+
16553
+
16554
+
16555
+
16556
+
16557
+
16558
+
16559
+
16560
+
16561
+
16562
+
16563
+
16564
+
16565
+
16566
+
16567
+
16568
+
16569
+
16570
+
16571
+
16572
+
16573
+
16574
+
16575
+
16576
+
16577
+
16578
+
16579
+
16580
+
16581
+
16582
+
16583
+
16584
+
16585
+
16586
+
16587
+
16588
+
16589
+
16590
+
16591
+
16592
+
16593
+
16594
+
16595
+
16596
+
16597
+
16598
+
16599
+
16600
+
16601
+
16602
+
16603
+
16604
+
16605
+
16606
+
16607
+
16608
+
16609
+
16610
+
16611
+
16612
+
16613
+
16614
+
16615
+
16616
+
16617
+
16618
+
16619
+
16620
+
16621
+
16622
+
16623
+
16624
+
16625
+
16626
+
16627
+
16628
+
16629
+
16630
+
16631
+
16632
+
16633
+
16634
+
16635
+
16636
+
16637
+
16638
+
16639
+
16640
+
16641
+
16642
+
16643
+
16644
+
16645
+
16646
+
16647
+
16648
+
16649
+
16650
+
16651
+
16652
+
16653
+
16654
+
16655
+
16656
+
16657
+
16658
+
16659
+
16660
+
16661
+
16662
+
16663
+
16664
+
16665
+
16666
+
16667
+
16668
+
16669
+
16670
+
16671
+
16672
+
16673
+
16674
+
16675
+
16676
+
16677
+
16678
+
16679
+
16680
+
16681
+
16682
+
16683
+
16684
+
16685
+
16686
+
16687
+
16688
+
16689
+
16690
+
16691
+
16692
+
16693
+
16694
+
16695
+
16696
+
16697
+
16698
+
16699
+
16700
+
16701
+
16702
+
16703
+
16704
+
16705
+
16706
+
16707
+
16708
+
16709
+
16710
+
16711
+
16712
+
16713
+
16714
+
16715
+
16716
+
16717
+
16718
+
16719
+
16720
+
16721
+
16722
+
16723
+
16724
+
16725
+
16726
+
16727
+
16728
+
16729
+
16730
+
16731
+
16732
+
16733
+
16734
+
16735
+
16736
+
16737
+
16738
+
16739
+
16740
+
16741
+
16742
+
16743
+
16744
+
16745
+
16746
+
16747
+
16748
+
16749
+
16750
+
16751
+
16752
+
16753
+
16754
+
16755
+
16756
+
16757
+
16758
+
16759
+
16760
+
16761
+
16762
+
16763
+
16764
+
16765
+
16766
+
16767
+
16768
+
16769
+
16770
+
16771
+
16772
+
16773
+
16774
+
16775
+
16776
+
16777
+
16778
+
16779
+
16780
+
16781
+
16782
+
16783
+
16784
+
16785
+
16786
+
16787
+
16788
+
16789
+
16790
+
16791
+
16792
+
16793
+
16794
+
16795
+
16796
+
16797
+
16798
+
16799
+
16800
+
16801
+
16802
+
16803
+
16804
+
16805
+
16806
+
16807
+
16808
+
16809
+
16810
+
16811
+
16812
+
16813
+
16814
+
16815
+
16816
+
16817
+
16818
+
16819
+
16820
+
16821
+
16822
+
16823
+
16824
+
16825
+
16826
+
16827
+
16828
+
16829
+
16830
+
16831
+
16832
+
16833
+
16834
+
16835
+
16836
+
16837
+
16838
+
16839
+
16840
+
16841
+
16842
+
16843
+
16844
+
16845
+
16846
+
16847
+
16848
+
16849
+
16850
+
16851
+
16852
+
16853
+
16854
+
16855
+
16856
+
16857
+
16858
+
16859
+
16860
+
16861
+
16862
+
16863
+
16864
+
16865
+
16866
+
16867
+
16868
+
16869
+
16870
+
16871
+
16872
+
16873
+
16874
+
16875
+
16876
+
16877
+
16878
+
16879
+
16880
+
16881
+
16882
+
16883
+
16884
+
16885
+
16886
+
16887
+
16888
+
16889
+
16890
+
16891
+
16892
+
16893
+
16894
+
16895
+
16896
+
16897
+
16898
+
16899
+
16900
+
16901
+
16902
+
16903
+
16904
+
16905
+
16906
+
16907
+
16908
+
16909
+
16910
+
16911
+
16912
+
16913
+ Step... (24000/50000 | Loss: 1.6508632898330688, Acc: 0.6671841740608215): 50%|█████████████▌ | 25000/50000 [9:54:56<11:05:00, 1.60s/it]
16914
+ Evaluating ...: 5%|████▍ | 6/130 [00:00<00:07, 15.90it/s]
16915
+ Step... (24500 | Loss: 1.7519614696502686, Learning Rate: 0.0003090909158345312)
16916
+
16917
+
16918
+
16919
+
16920
+
16921
+
16922
+
16923
+
16924
+
16925
+
16926
+
16927
+
16928
+ [12:03:19] - INFO - __main__ - Saving checkpoint at 25000 steps█████████████████████████████████████████████████████| 130/130 [00:21<00:00, 4.60it/s]
16929
+ All Flax model weights were used when initializing RobertaForMaskedLM.
16930
+ Some weights of RobertaForMaskedLM were not initialized from the Flax model and are newly initialized: ['lm_head.decoder.weight', 'roberta.embeddings.position_ids', 'lm_head.decoder.bias']
16931
+ You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
16932
+
16933
+
16934
+
16935
+
16936
+
16937
+
16938
+
16939
+
16940
+
16941
+
16942
+
16943
+
16944
+
16945
+
16946
+
16947
+
16948
+
16949
+
16950
+
16951
+
16952
+
16953
+
16954
+
16955
+
16956
+
16957
+
16958
+
16959
+
16960
+
16961
+
16962
+
16963
+
16964
+
16965
+
16966
+
16967
+
16968
+
16969
+
16970
+
16971
+
16972
+
16973
+
16974
+
16975
+
16976
+
16977
+
16978
+
16979
+
16980
+
16981
+
16982
+
16983
+
16984
+
16985
+
16986
+
16987
+
16988
+
16989
+
16990
+
16991
+
16992
+
16993
+
16994
+
16995
+
16996
+
16997
+
16998
+
16999
+
17000
+
17001
+
17002
+
17003
+
17004
+
17005
+
17006
+
17007
+
17008
+
17009
+
17010
+
17011
+
17012
+
17013
+
17014
+
17015
+
17016
+
17017
+
17018
+
17019
+
17020
+
17021
+
17022
+
17023
+
17024
+
17025
+
17026
+
17027
+
17028
+
17029
+
17030
+
17031
+
17032
+
17033
+
17034
+
17035
+
17036
+
17037
+
17038
+
17039
+
17040
+
17041
+
17042
+
17043
+
17044
+
17045
+
17046
+
17047
+
17048
+
17049
+
17050
+
17051
+
17052
+
17053
+
17054
+
17055
+
17056
+
17057
+
17058
+
17059
+
17060
+
17061
+
17062
+
17063
+
17064
+
17065
+
17066
+
17067
+
17068
+
17069
+
17070
+
17071
+
17072
+
17073
+
17074
+
17075
+
17076
+
17077
+
17078
+
17079
+
17080
+
17081
+
17082
+
17083
+
17084
+
17085
+
17086
+
17087
+
17088
+
17089
+
17090
+
17091
+
17092
+
17093
+
17094
+
17095
+
17096
+
17097
+
17098
+
17099
+
17100
+
17101
+
17102
+
17103
+
17104
+
17105
+
17106
+
17107
+
17108
+
17109
+
17110
+
17111
+
17112
+
17113
+
17114
+
17115
+
17116
+
17117
+
17118
+
17119
+
17120
+
17121
+
17122
+
17123
+
17124
+
17125
+
17126
+
17127
+
17128
+
17129
+
17130
+
17131
+
17132
+
17133
+
17134
+
17135
+
17136
+
17137
+
17138
+
17139
+
17140
+
17141
+
17142
+
17143
+
17144
+
17145
+
17146
+
17147
+
17148
+
17149
+
17150
+
17151
+
17152
+
17153
+
17154
+
17155
+
17156
+
17157
+
17158
+
17159
+
17160
+
17161
+
17162
+
17163
+
17164
+
17165
+
17166
+
17167
+
17168
+
17169
+
17170
+
17171
+
17172
+
17173
+
17174
+
17175
+
17176
+
17177
+
17178
+
17179
+
17180
+
17181
+
17182
+
17183
+
17184
+
17185
+
17186
+
17187
+
17188
+
17189
+
17190
+
17191
+
17192
+
17193
+
17194
+
17195
+
17196
+
17197
+
17198
+
17199
+
17200
+
17201
+
17202
+
17203
+
17204
+
17205
+
17206
+
17207
+
17208
+
17209
+
17210
+
17211
+
17212
+
17213
+
17214
+
17215
+
17216
+
17217
+
17218
+
17219
+
17220
+
17221
+
17222
+
17223
+
17224
+
17225
+
17226
+
17227
+
17228
+
17229
+
17230
+
17231
+
17232
+
17233
+
17234
+
17235
+
17236
+
17237
+
17238
+
17239
+
17240
+
17241
+
17242
+
17243
+
17244
+
17245
+
17246
+
17247
+
17248
+
17249
+
17250
+
17251
+
17252
+
17253
+
17254
+
17255
+
17256
+
17257
+
17258
+
17259
+
17260
+
17261
+
17262
+
17263
+
17264
+
17265
+
17266
+
17267
+
17268
+
17269
+
17270
+
17271
+
17272
+
17273
+
17274
+
17275
+
17276
+
17277
+
17278
+
17279
+
17280
+
17281
+
17282
+
17283
+
17284
+
17285
+
17286
+
17287
+
17288
+
17289
+
17290
+
17291
+
17292
+
17293
+
17294
+
17295
+
17296
+
17297
+
17298
+
17299
+
17300
+
17301
+
17302
+
17303
+
17304
+
17305
+
17306
+
17307
+
17308
+
17309
+
17310
+
17311
+
17312
+
17313
+
17314
+
17315
+
17316
+
17317
+
17318
+
17319
+
17320
+
17321
+
17322
+
17323
+
17324
+
17325
+
17326
+
17327
+
17328
+
17329
+
17330
+
17331
+
17332
+
17333
+
17334
+
17335
+
17336
+
17337
+
17338
+
17339
+
17340
+
17341
+
17342
+
17343
+
17344
+
17345
+
17346
+
17347
+
17348
+
17349
+
17350
+
17351
+
17352
+
17353
+
17354
+
17355
+
17356
+
17357
+
17358
+
17359
+
17360
+
17361
+
17362
+
17363
+
17364
+
17365
+
17366
+
17367
+
17368
+
17369
+
17370
+
17371
+
17372
+
17373
+
17374
+
17375
+
17376
+
17377
+
17378
+
17379
+
17380
+
17381
+
17382
+
17383
+
17384
+
17385
+
17386
+
17387
+
17388
+
17389
+
17390
+
17391
+
17392
+
17393
+
17394
+
17395
+
17396
+
17397
+
17398
+
17399
+
17400
+
17401
+
17402
+
17403
+
17404
+
17405
+
17406
+
17407
+
17408
+
17409
+
17410
+
17411
+
17412
+
17413
+
17414
+
17415
+
17416
+
17417
+
17418
+
17419
+
17420
+
17421
+
17422
+
17423
+
17424
+
17425
+
17426
+
17427
+
17428
+
17429
+
17430
+
17431
+
17432
+
17433
+
17434
+
17435
+
17436
+
17437
+
17438
+
17439
+
17440
+
17441
+
17442
+
17443
+
17444
+
17445
+
17446
+
17447
+
17448
+
17449
+
17450
+
17451
+
17452
+
17453
+
17454
+
17455
+
17456
+
17457
+
17458
+
17459
+
17460
+
17461
+
17462
+
17463
+
17464
+
17465
+
17466
+
17467
+
17468
+
17469
+
17470
+
17471
+
17472
+
17473
+
17474
+
17475
+
17476
+
17477
+
17478
+
17479
+
17480
+
17481
+
17482
+
17483
+
17484
+
17485
+
17486
+
17487
+
17488
+
17489
+
17490
+
17491
+
17492
+
17493
+
17494
+
17495
+
17496
+
17497
+
17498
+
17499
+
17500
+
17501
+
17502
+
17503
+
17504
+
17505
+
17506
+
17507
+
17508
+
17509
+
17510
+
17511
+
17512
+
17513
+
17514
+
17515
+
17516
+
17517
+
17518
+
17519
+
17520
+
17521
+
17522
+
17523
+
17524
+
17525
+
17526
+
17527
+
17528
+
17529
+
17530
+
17531
+
17532
+
17533
+
17534
+
17535
+
17536
+
17537
+
17538
+
17539
+
17540
+
17541
+
17542
+
17543
+
17544
+
17545
+
17546
+
17547
+
17548
+
17549
+
17550
+
17551
+
17552
+
17553
+
17554
+
17555
+
17556
+
17557
+
17558
+
17559
+
17560
+
17561
+
17562
+
17563
+
17564
+
17565
+
17566
+
17567
+
17568
+
17569
+
17570
+
17571
+
17572
+
17573
+
17574
+
17575
+
17576
+
17577
+
17578
+
17579
+
17580
+
17581
+
17582
+
17583
+
17584
+
17585
+
17586
+
17587
+
17588
+
17589
+
17590
+
17591
+
17592
+
17593
+
17594
+
17595
+
17596
+
17597
+
17598
+
17599
+
17600
+
17601
+
17602
+
17603
+
17604
+
17605
+
17606
+
17607
+
17608
+
17609
+
17610
+
17611
+
17612
+
17613
+ Step... (25000/50000 | Loss: 1.6436606645584106, Acc: 0.668701171875): 52%|████████████████ | 26000/50000 [10:20:04<8:52:22, 1.33s/it]
17614
+ Step... (25500 | Loss: 1.6520822048187256, Learning Rate: 0.0002969697234220803)
17615
+ Step... (26000 | Loss: 1.7167686223983765, Learning Rate: 0.0002909091126639396)
17616
+
17617
+
17618
+
17619
+
17620
+
17621
+
17622
+
17623
+
17624
+
17625
+
17626
+
17627
+ [12:28:28] - INFO - __main__ - Saving checkpoint at 26000 steps█████████████████████████████████████████████████████| 130/130 [00:21<00:00, 4.60it/s]
17628
+ All Flax model weights were used when initializing RobertaForMaskedLM.
17629
+ Some weights of RobertaForMaskedLM were not initialized from the Flax model and are newly initialized: ['lm_head.decoder.weight', 'roberta.embeddings.position_ids', 'lm_head.decoder.bias']
17630
+ You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
17631
+
17632
+
17633
+
17634
+
17635
+
17636
+
17637
+
17638
+
17639
+
17640
+
17641
+
17642
+
17643
+
17644
+
17645
+
17646
+
17647
+
17648
+
17649
+
17650
+
17651
+
17652
+
17653
+
17654
+
17655
+
17656
+
17657
+
17658
+
17659
+
17660
+
17661
+
17662
+
17663
+
17664
+
17665
+
17666
+
17667
+
17668
+
17669
+
17670
+
17671
+
17672
+
17673
+
17674
+
17675
+
17676
+
17677
+
17678
+
17679
+
17680
+
17681
+
17682
+
17683
+
17684
+
17685
+
17686
+
17687
+
17688
+
17689
+
17690
+
17691
+
17692
+
17693
+
17694
+
17695
+
17696
+
17697
+
17698
+
17699
+
17700
+
17701
+
17702
+
17703
+
17704
+
17705
+
17706
+
17707
+
17708
+
17709
+
17710
+
17711
+
17712
+
17713
+
17714
+
17715
+
17716
+
17717
+
17718
+
17719
+
17720
+
17721
+
17722
+
17723
+
17724
+
17725
+
17726
+
17727
+
17728
+
17729
+
17730
+
17731
+
17732
+
17733
+
17734
+
17735
+
17736
+
17737
+
17738
+
17739
+
17740
+
17741
+
17742
+
17743
+
17744
+
17745
+
17746
+
17747
+
17748
+
17749
+
17750
+
17751
+
17752
+
17753
+
17754
+
17755
+
17756
+
17757
+
17758
+
17759
+
17760
+
17761
+
17762
+
17763
+
17764
+
17765
+
17766
+
17767
+
17768
+
17769
+
17770
+
17771
+
17772
+
17773
+
17774
+
17775
+
17776
+
17777
+
17778
+
17779
+
17780
+
17781
+
17782
+
17783
+
17784
+
17785
+
17786
+
17787
+
17788
+
17789
+
17790
+
17791
+
17792
+
17793
+
17794
+
17795
+
17796
+
17797
+
17798
+
17799
+
17800
+
17801
+
17802
+
17803
+
17804
+
17805
+
17806
+
17807
+
17808
+
17809
+
17810
+
17811
+
17812
+
17813
+
17814
+
17815
+
17816
+
17817
+
17818
+
17819
+
17820
+
17821
+
17822
+
17823
+
17824
+
17825
+
17826
+
17827
+
17828
+
17829
+
17830
+
17831
+
17832
+
17833
+
17834
+
17835
+
17836
+
17837
+
17838
+
17839
+
17840
+
17841
+
17842
+
17843
+
17844
+
17845
+
17846
+
17847
+
17848
+
17849
+
17850
+
17851
+
17852
+
17853
+
17854
+
17855
+
17856
+
17857
+
17858
+
17859
+
17860
+
17861
+
17862
+
17863
+
17864
+
17865
+
17866
+
17867
+
17868
+
17869
+
17870
+
17871
+
17872
+
17873
+
17874
+
17875
+
17876
+
17877
+
17878
+
17879
+
17880
+
17881
+
17882
+
17883
+
17884
+
17885
+
17886
+
17887
+
17888
+
17889
+
17890
+
17891
+
17892
+
17893
+
17894
+
17895
+
17896
+
17897
+
17898
+
17899
+
17900
+
17901
+
17902
+
17903
+
17904
+
17905
+
17906
+
17907
+
17908
+
17909
+
17910
+
17911
+
17912
+
17913
+
17914
+
17915
+
17916
+
17917
+
17918
+
17919
+
17920
+
17921
+
17922
+
17923
+
17924
+
17925
+
17926
+
17927
+
17928
+
17929
+
17930
+
17931
+
17932
+
17933
+
17934
+
17935
+
17936
+
17937
+
17938
+
17939
+
17940
+
17941
+
17942
+
17943
+
17944
+
17945
+
17946
+
17947
+
17948
+
17949
+
17950
+
17951
+
17952
+
17953
+
17954
+
17955
+
17956
+
17957
+
17958
+
17959
+
17960
+
17961
+
17962
+
17963
+
17964
+
17965
+
17966
+
17967
+
17968
+
17969
+
17970
+
17971
+
17972
+
17973
+
17974
+
17975
+
17976
+
17977
+
17978
+
17979
+
17980
+
17981
+
17982
+
17983
+
17984
+
17985
+
17986
+
17987
+
17988
+
17989
+
17990
+
17991
+
17992
+
17993
+
17994
+
17995
+
17996
+
17997
+
17998
+
17999
+
18000
+
18001
+
18002
+
18003
+
18004
+
18005
+
18006
+
18007
+
18008
+
18009
+
18010
+
18011
+
18012
+
18013
+
18014
+
18015
+
18016
+
18017
+
18018
+
18019
+
18020
+
18021
+
18022
+
18023
+
18024
+
18025
+
18026
+
18027
+
18028
+
18029
+
18030
+
18031
+
18032
+
18033
+
18034
+
18035
+
18036
+
18037
+
18038
+
18039
+
18040
+
18041
+
18042
+
18043
+
18044
+
18045
+
18046
+
18047
+
18048
+
18049
+
18050
+
18051
+
18052
+
18053
+
18054
 
18055
 
18056
 
wandb/run-20210726_001233-17u6inbn/files/wandb-summary.json CHANGED
@@ -1 +1 @@
1
- {"global_step": 24000, "_timestamp": 1627299487.452405, "train_time": 1156106.125, "train_learning_rate": 0.00031515152659267187, "_step": 47856, "train_loss": 1.7166345119476318, "eval_accuracy": 0.6663545966148376, "eval_loss": 1.6572293043136597}
 
1
+ {"global_step": 26500, "_timestamp": 1627303266.586647, "train_time": 1372972.75, "train_learning_rate": 0.0002848485019057989, "_step": 52841, "train_loss": 1.730733036994934, "eval_accuracy": 0.6691190600395203, "eval_loss": 1.6362030506134033}
wandb/run-20210726_001233-17u6inbn/logs/debug-internal.log CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e82989e4b19c6c0abd610b0181219b8926bc8d5e7d84c1812150b24b6b6a4d6e
3
- size 18951993
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45d92a70b4478861c91b2901d46e09c9d061fd6a02da2c0ff8ca5335c33cfde8
3
+ size 20913336
wandb/run-20210726_001233-17u6inbn/run-17u6inbn.wandb CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0c32d64082b6ac9a729c131c88cc2d56813251ca3d7cc69eb10cf688204a79ff
3
- size 9437234
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1343d0ef55818f25efb5ed042e2cf55bb6c6f3440aba38c3b2f266129989e652
3
+ size 10444427