rootxhacker commited on
Commit
67c6d00
·
verified ·
1 Parent(s): 7686112

Training in progress, step 37000, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:00da868c16026f8b91fd852d53e5b1e8b84bd1c76aefa3ce6543ff5082c8a8c1
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c033c68f7f85112dd0424df6a6c069e43bad62023f4f8046088b5f303d66757d
3
  size 36730224
last-checkpoint/ar_diffusion_info.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:04e0ccc68b635616af9e0dbead055411d758f4619095d018f21601b71ae1dad9
3
  size 1544
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:801dfb4f56aec672f8a7307f1b9530439cf1a3e5150aec9ed045fd57278ee3b1
3
  size 1544
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:81484f3b03bf089df3397a87ac9d9ffd5e1f946de717b51932df743ee8d9f915
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:350f65f9d77fda04ddb0fb82bddcf712e278846103610b6be4eb0e5b91d1deaa
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5d10178046632fe47bb7008cd9e1a9974e0ea49262160950299e919075603393
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa1f6feef07e5b1d8fea95579cd14290f66870371d42761bb6d273aa098f1986
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:81382d5bc209ea226a1381f99c5863a7a3af33f40b32204afd2996013b6f09c3
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08aff465a459746ab9dbe30f49ac4e0ea77e8a027d139d1bf09d3f39d9942cdb
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0cf91a495853427b5f841ab1be3c35766ad0edbc12ff06e2c15160a47811ed22
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0bc995789ac7ace85eec5527f15f9a82c9f1388944ba2d5baa678f54ce3d8943
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 34000,
3
- "best_metric": 0.987713634967804,
4
  "best_model_checkpoint": "./ar-diffusion-checkpoints-fixed/checkpoint-34000",
5
- "epoch": 2.8074763479732328,
6
  "eval_steps": 250,
7
- "global_step": 36500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -6286,6 +6286,92 @@
6286
  "eval_samples_per_second": 57.54,
6287
  "eval_steps_per_second": 14.385,
6288
  "step": 36500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6289
  }
6290
  ],
6291
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 36750,
3
+ "best_metric": 0.9847651720046997,
4
  "best_model_checkpoint": "./ar-diffusion-checkpoints-fixed/checkpoint-34000",
5
+ "epoch": 2.845934928082455,
6
  "eval_steps": 250,
7
+ "global_step": 37000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
6286
  "eval_samples_per_second": 57.54,
6287
  "eval_steps_per_second": 14.385,
6288
  "step": 36500
6289
+ },
6290
+ {
6291
+ "epoch": 2.811322205984155,
6292
+ "grad_norm": 0.9856983423233032,
6293
+ "learning_rate": 1.2824974677297874e-05,
6294
+ "loss": 0.9923,
6295
+ "step": 36550
6296
+ },
6297
+ {
6298
+ "epoch": 2.8151680639950776,
6299
+ "grad_norm": 1.2112038135528564,
6300
+ "learning_rate": 1.2565254655481392e-05,
6301
+ "loss": 0.9388,
6302
+ "step": 36600
6303
+ },
6304
+ {
6305
+ "epoch": 2.8190139220059995,
6306
+ "grad_norm": 1.1824342012405396,
6307
+ "learning_rate": 1.230553463366491e-05,
6308
+ "loss": 0.984,
6309
+ "step": 36650
6310
+ },
6311
+ {
6312
+ "epoch": 2.8228597800169215,
6313
+ "grad_norm": 1.3278725147247314,
6314
+ "learning_rate": 1.204581461184843e-05,
6315
+ "loss": 0.9835,
6316
+ "step": 36700
6317
+ },
6318
+ {
6319
+ "epoch": 2.826705638027844,
6320
+ "grad_norm": 1.4297553300857544,
6321
+ "learning_rate": 1.1786094590031946e-05,
6322
+ "loss": 0.9999,
6323
+ "step": 36750
6324
+ },
6325
+ {
6326
+ "epoch": 2.826705638027844,
6327
+ "eval_loss": 0.9847651720046997,
6328
+ "eval_runtime": 17.5097,
6329
+ "eval_samples_per_second": 57.111,
6330
+ "eval_steps_per_second": 14.278,
6331
+ "step": 36750
6332
+ },
6333
+ {
6334
+ "epoch": 2.8305514960387663,
6335
+ "grad_norm": 1.1944117546081543,
6336
+ "learning_rate": 1.1526374568215465e-05,
6337
+ "loss": 0.9758,
6338
+ "step": 36800
6339
+ },
6340
+ {
6341
+ "epoch": 2.8343973540496883,
6342
+ "grad_norm": 1.1629287004470825,
6343
+ "learning_rate": 1.1266654546398983e-05,
6344
+ "loss": 0.988,
6345
+ "step": 36850
6346
+ },
6347
+ {
6348
+ "epoch": 2.8382432120606107,
6349
+ "grad_norm": 1.598382592201233,
6350
+ "learning_rate": 1.1006934524582501e-05,
6351
+ "loss": 0.9679,
6352
+ "step": 36900
6353
+ },
6354
+ {
6355
+ "epoch": 2.842089070071533,
6356
+ "grad_norm": 0.9534172415733337,
6357
+ "learning_rate": 1.0747214502766019e-05,
6358
+ "loss": 0.9921,
6359
+ "step": 36950
6360
+ },
6361
+ {
6362
+ "epoch": 2.845934928082455,
6363
+ "grad_norm": 1.5716655254364014,
6364
+ "learning_rate": 1.0487494480949537e-05,
6365
+ "loss": 0.9898,
6366
+ "step": 37000
6367
+ },
6368
+ {
6369
+ "epoch": 2.845934928082455,
6370
+ "eval_loss": 0.9878412485122681,
6371
+ "eval_runtime": 17.2808,
6372
+ "eval_samples_per_second": 57.868,
6373
+ "eval_steps_per_second": 14.467,
6374
+ "step": 37000
6375
  }
6376
  ],
6377
  "logging_steps": 50,