shuheng commited on
Commit
21c416c
verified
1 Parent(s): 14f93ce

End of training

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ eval_nbest_predictions.json filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -4,6 +4,8 @@ license: apache-2.0
4
  base_model: albert/albert-xlarge-v2
5
  tags:
6
  - generated_from_trainer
 
 
7
  model-index:
8
  - name: trivia_albert_xl_finetuned
9
  results: []
@@ -14,7 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  # trivia_albert_xl_finetuned
16
 
17
- This model is a fine-tuned version of [albert/albert-xlarge-v2](https://huggingface.co/albert/albert-xlarge-v2) on an unknown dataset.
18
 
19
  ## Model description
20
 
 
4
  base_model: albert/albert-xlarge-v2
5
  tags:
6
  - generated_from_trainer
7
+ datasets:
8
+ - TimoImhof/TriviaQA-in-SQuAD-format
9
  model-index:
10
  - name: trivia_albert_xl_finetuned
11
  results: []
 
16
 
17
  # trivia_albert_xl_finetuned
18
 
19
+ This model is a fine-tuned version of [albert/albert-xlarge-v2](https://huggingface.co/albert/albert-xlarge-v2) on the TimoImhof/TriviaQA-in-SQuAD-format dataset.
20
 
21
  ## Model description
22
 
all_results.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 6.0,
3
+ "eval_exact_match": 75.40663630448927,
4
+ "eval_f1": 79.50348921131032,
5
+ "eval_runtime": 109.1736,
6
+ "eval_samples": 3398,
7
+ "eval_samples_per_second": 31.125,
8
+ "eval_steps_per_second": 3.893,
9
+ "total_flos": 9479673213373440.0,
10
+ "train_loss": 1.1461569756831762,
11
+ "train_runtime": 6870.8994,
12
+ "train_samples": 13545,
13
+ "train_samples_per_second": 11.828,
14
+ "train_steps_per_second": 0.37
15
+ }
eval_nbest_predictions.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d2ece278e581fdacaa19d135f96eab4b9c6dc621a7c2155251fd021d71da4958
3
+ size 13712303
eval_predictions.json ADDED
The diff for this file is too large to render. See raw diff
 
eval_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 6.0,
3
+ "eval_exact_match": 75.40663630448927,
4
+ "eval_f1": 79.50348921131032,
5
+ "eval_runtime": 109.1736,
6
+ "eval_samples": 3398,
7
+ "eval_samples_per_second": 31.125,
8
+ "eval_steps_per_second": 3.893
9
+ }
runs/Feb02_05-11-35_xgpi2/events.out.tfevents.1738451324.xgpi2.2490936.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:abea0d3f3c9a5db7ef40c45fa6e0c8138baa89c418c793ac4c2e7d3c757c80ac
3
+ size 412
train_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 6.0,
3
+ "total_flos": 9479673213373440.0,
4
+ "train_loss": 1.1461569756831762,
5
+ "train_runtime": 6870.8994,
6
+ "train_samples": 13545,
7
+ "train_samples_per_second": 11.828,
8
+ "train_steps_per_second": 0.37
9
+ }
trainer_state.json ADDED
@@ -0,0 +1,77 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 6.0,
5
+ "eval_steps": 500,
6
+ "global_step": 2544,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.179245283018868,
13
+ "grad_norm": 28.103540420532227,
14
+ "learning_rate": 2.410377358490566e-05,
15
+ "loss": 2.2352,
16
+ "step": 500
17
+ },
18
+ {
19
+ "epoch": 2.358490566037736,
20
+ "grad_norm": 16.953514099121094,
21
+ "learning_rate": 1.8207547169811322e-05,
22
+ "loss": 1.5029,
23
+ "step": 1000
24
+ },
25
+ {
26
+ "epoch": 3.5377358490566038,
27
+ "grad_norm": 31.308820724487305,
28
+ "learning_rate": 1.2311320754716983e-05,
29
+ "loss": 1.1127,
30
+ "step": 1500
31
+ },
32
+ {
33
+ "epoch": 4.716981132075472,
34
+ "grad_norm": 26.33289337158203,
35
+ "learning_rate": 6.415094339622642e-06,
36
+ "loss": 0.7052,
37
+ "step": 2000
38
+ },
39
+ {
40
+ "epoch": 5.89622641509434,
41
+ "grad_norm": 106.1405258178711,
42
+ "learning_rate": 5.188679245283019e-07,
43
+ "loss": 0.2618,
44
+ "step": 2500
45
+ },
46
+ {
47
+ "epoch": 6.0,
48
+ "step": 2544,
49
+ "total_flos": 9479673213373440.0,
50
+ "train_loss": 1.1461569756831762,
51
+ "train_runtime": 6870.8994,
52
+ "train_samples_per_second": 11.828,
53
+ "train_steps_per_second": 0.37
54
+ }
55
+ ],
56
+ "logging_steps": 500,
57
+ "max_steps": 2544,
58
+ "num_input_tokens_seen": 0,
59
+ "num_train_epochs": 6,
60
+ "save_steps": 500,
61
+ "stateful_callbacks": {
62
+ "TrainerControl": {
63
+ "args": {
64
+ "should_epoch_stop": false,
65
+ "should_evaluate": false,
66
+ "should_log": false,
67
+ "should_save": true,
68
+ "should_training_stop": true
69
+ },
70
+ "attributes": {}
71
+ }
72
+ },
73
+ "total_flos": 9479673213373440.0,
74
+ "train_batch_size": 32,
75
+ "trial_name": null,
76
+ "trial_params": null
77
+ }