rakhman-llm commited on
Commit
dd7d1b8
·
verified ·
1 Parent(s): f53495d

Training in progress, step 38500, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9b01c6e320f3b8ce398fbd50bb34cd5deb4150ad4b0e09a91d304d07ef6a1d44
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61083c956d931fcb4b0a0b08aca90fdf494df1d92e15a552d2a52a3b02538c7a
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:37f54042d6d8e987a2001e8f8b69e12f7b5e1be6322534bf71acae8f44c1d295
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee11215041eb0186976ef4b18acc8c3b3085b8278e5c482e739c8198da6b7b9a
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:61ae2baae6f9dfd7ca89f4f0f5818402f18a3e15e4581cd68734c6a76f2a7030
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9244e8a4fedebd96fe9477d928481702bf2954643113838dc31cc9779fe5ad3
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e81910e79bb87f0b46a0d2aa6ab0730eb92717d10eebce124863a9bc14f71612
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1bfe4e613c5d63b9b8661f26b2521341654055e778c8bc3f3f8cced7a4c2ef7b
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 2.533333333333333,
5
  "eval_steps": 500,
6
- "global_step": 38000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -2683,6 +2683,41 @@
2683
  "learning_rate": 3.117333333333333e-06,
2684
  "loss": 0.0477,
2685
  "step": 38000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2686
  }
2687
  ],
2688
  "logging_steps": 100,
@@ -2702,7 +2737,7 @@
2702
  "attributes": {}
2703
  }
2704
  },
2705
- "total_flos": 9.256159936512e+16,
2706
  "train_batch_size": 4,
2707
  "trial_name": null,
2708
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 2.5666666666666664,
5
  "eval_steps": 500,
6
+ "global_step": 38500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
2683
  "learning_rate": 3.117333333333333e-06,
2684
  "loss": 0.0477,
2685
  "step": 38000
2686
+ },
2687
+ {
2688
+ "epoch": 2.54,
2689
+ "grad_norm": 0.23089808225631714,
2690
+ "learning_rate": 3.0728888888888893e-06,
2691
+ "loss": 0.0544,
2692
+ "step": 38100
2693
+ },
2694
+ {
2695
+ "epoch": 2.546666666666667,
2696
+ "grad_norm": 0.2545056939125061,
2697
+ "learning_rate": 3.028444444444445e-06,
2698
+ "loss": 0.049,
2699
+ "step": 38200
2700
+ },
2701
+ {
2702
+ "epoch": 2.5533333333333332,
2703
+ "grad_norm": 0.2878418564796448,
2704
+ "learning_rate": 2.984e-06,
2705
+ "loss": 0.0506,
2706
+ "step": 38300
2707
+ },
2708
+ {
2709
+ "epoch": 2.56,
2710
+ "grad_norm": 0.2358868569135666,
2711
+ "learning_rate": 2.9395555555555562e-06,
2712
+ "loss": 0.0539,
2713
+ "step": 38400
2714
+ },
2715
+ {
2716
+ "epoch": 2.5666666666666664,
2717
+ "grad_norm": 0.20847082138061523,
2718
+ "learning_rate": 2.8951111111111114e-06,
2719
+ "loss": 0.0549,
2720
+ "step": 38500
2721
  }
2722
  ],
2723
  "logging_steps": 100,
 
2737
  "attributes": {}
2738
  }
2739
  },
2740
+ "total_flos": 9.377951514624e+16,
2741
  "train_batch_size": 4,
2742
  "trial_name": null,
2743
  "trial_params": null