yongzx commited on
Commit
bc327b5
·
1 Parent(s): ccdf555

rm --cached *

Browse files
.gitattributes DELETED
@@ -1,33 +0,0 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bz2 filter=lfs diff=lfs merge=lfs -text
5
- *.ftz filter=lfs diff=lfs merge=lfs -text
6
- *.gz filter=lfs diff=lfs merge=lfs -text
7
- *.h5 filter=lfs diff=lfs merge=lfs -text
8
- *.joblib filter=lfs diff=lfs merge=lfs -text
9
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
10
- *.mlmodel filter=lfs diff=lfs merge=lfs -text
11
- *.model filter=lfs diff=lfs merge=lfs -text
12
- *.msgpack filter=lfs diff=lfs merge=lfs -text
13
- *.npy filter=lfs diff=lfs merge=lfs -text
14
- *.npz filter=lfs diff=lfs merge=lfs -text
15
- *.onnx filter=lfs diff=lfs merge=lfs -text
16
- *.ot filter=lfs diff=lfs merge=lfs -text
17
- *.parquet filter=lfs diff=lfs merge=lfs -text
18
- *.pb filter=lfs diff=lfs merge=lfs -text
19
- *.pickle filter=lfs diff=lfs merge=lfs -text
20
- *.pkl filter=lfs diff=lfs merge=lfs -text
21
- *.pt filter=lfs diff=lfs merge=lfs -text
22
- *.pth filter=lfs diff=lfs merge=lfs -text
23
- *.rar filter=lfs diff=lfs merge=lfs -text
24
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
25
- *.tar.* filter=lfs diff=lfs merge=lfs -text
26
- *.tflite filter=lfs diff=lfs merge=lfs -text
27
- *.tgz filter=lfs diff=lfs merge=lfs -text
28
- *.wasm filter=lfs diff=lfs merge=lfs -text
29
- *.xz filter=lfs diff=lfs merge=lfs -text
30
- *.zip filter=lfs diff=lfs merge=lfs -text
31
- *.zst filter=lfs diff=lfs merge=lfs -text
32
- *tfevents* filter=lfs diff=lfs merge=lfs -text
33
- tokenizer.json filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
.gitignore DELETED
@@ -1,15 +0,0 @@
1
- checkpoint-*/
2
- */pilot_*/
3
- pilot_*/
4
- checkpoint-*/
5
- */pilot_*/
6
- pilot_*/
7
- checkpoint-*/
8
- */pilot_*/
9
- pilot_*/
10
- checkpoint-*/
11
- */pilot_*/
12
- pilot_*/
13
- checkpoint-*/
14
- */pilot_*/
15
- pilot_*/
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
all_results.json DELETED
@@ -1,14 +0,0 @@
1
- {
2
- "epoch": 2.13,
3
- "eval_loss": 2.2437143325805664,
4
- "eval_runtime": 482.5676,
5
- "eval_samples": 4608,
6
- "eval_samples_per_second": 9.549,
7
- "eval_steps_per_second": 4.774,
8
- "perplexity": 9.428286120634498,
9
- "train_loss": 2.3864154296875,
10
- "train_runtime": 60699.5161,
11
- "train_samples": 94080,
12
- "train_samples_per_second": 3.295,
13
- "train_steps_per_second": 0.412
14
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
config.json DELETED
@@ -1,39 +0,0 @@
1
- {
2
- "_name_or_path": "bigscience/bloom-350m",
3
- "adapters": {
4
- "adapters": {},
5
- "config_map": {},
6
- "fusion_config_map": {},
7
- "fusions": {}
8
- },
9
- "apply_residual_connection_post_layernorm": false,
10
- "architectures": [
11
- "BloomForCausalLM"
12
- ],
13
- "attention_dropout": 0.0,
14
- "attention_softmax_in_fp32": true,
15
- "bias_dropout_fusion": true,
16
- "bos_token_id": 1,
17
- "eos_token_id": 2,
18
- "hidden_dropout": 0.0,
19
- "hidden_size": 1024,
20
- "initializer_range": 0.02,
21
- "layer_norm_epsilon": 1e-05,
22
- "masked_softmax_fusion": true,
23
- "model_type": "bloom",
24
- "n_head": 16,
25
- "n_inner": null,
26
- "n_layer": 24,
27
- "offset_alibi": 100,
28
- "pad_token_id": 3,
29
- "pretraining_tp": 1,
30
- "seq_length": 2048,
31
- "skip_bias_add": true,
32
- "skip_bias_add_qkv": false,
33
- "slow_but_exact": false,
34
- "torch_dtype": "float32",
35
- "transformers_version": "4.20.0.dev0",
36
- "unk_token_id": 0,
37
- "use_cache": true,
38
- "vocab_size": 250880
39
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
eval_results.json DELETED
@@ -1,9 +0,0 @@
1
- {
2
- "epoch": 2.13,
3
- "eval_loss": 2.2437143325805664,
4
- "eval_runtime": 482.5676,
5
- "eval_samples": 4608,
6
- "eval_samples_per_second": 9.549,
7
- "eval_steps_per_second": 4.774,
8
- "perplexity": 9.428286120634498
9
- }
 
 
 
 
 
 
 
 
 
 
pytorch_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:cef1be9a44e7ab905acaed1238738e494279ad43f8d0412d49dff23d3c7aed63
3
- size 2236955191
 
 
 
 
special_tokens_map.json DELETED
@@ -1 +0,0 @@
1
- {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>"}
 
 
tatoeba-az-en-results.txt DELETED
@@ -1,7 +0,0 @@
1
- ==================================================
2
- Tatoeba Results (200 pairs of az-en)
3
- ==================================================
4
- Model: /users/zyong2/data/zyong2/bigscience/data/processed/024/bloom-350m_az_fish_100000samples_-1vocab_original-frozen
5
- [tensor(0.0150)]
6
- 1.50 ± 0.00
7
- ==================================================
 
 
 
 
 
 
 
 
tokenizer.json DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:8f6efc66e73f1fd69da4f436e48befb519fdff3fe18910850c1d41bd862293a5
3
- size 14500443
 
 
 
 
tokenizer_config.json DELETED
@@ -1 +0,0 @@
1
- {"unk_token": "<unk>", "bos_token": "<s>", "eos_token": "</s>", "pad_token": "<pad>", "add_prefix_space": false, "name_or_path": "bigscience/bloom-350m", "special_tokens_map_file": null, "padding_side": "left", "tokenizer_class": "BloomTokenizer"}
 
 
train_results.json DELETED
@@ -1,8 +0,0 @@
1
- {
2
- "epoch": 2.13,
3
- "train_loss": 2.3864154296875,
4
- "train_runtime": 60699.5161,
5
- "train_samples": 94080,
6
- "train_samples_per_second": 3.295,
7
- "train_steps_per_second": 0.412
8
- }
 
 
 
 
 
 
 
 
 
trainer_state.json DELETED
@@ -1,125 +0,0 @@
1
- {
2
- "best_metric": 2.2437143325805664,
3
- "best_model_checkpoint": "/users/zyong2/data/zyong2/bigscience/data/processed/024/bloom-350m_az_fish_100000samples_-1vocab_original-frozen/checkpoint-25000",
4
- "epoch": 2.1258503401360542,
5
- "global_step": 25000,
6
- "is_hyper_param_search": false,
7
- "is_local_process_zero": true,
8
- "is_world_process_zero": true,
9
- "log_history": [
10
- {
11
- "epoch": 0.21,
12
- "learning_rate": 9e-05,
13
- "loss": 2.9924,
14
- "step": 2500
15
- },
16
- {
17
- "epoch": 0.43,
18
- "learning_rate": 8e-05,
19
- "loss": 2.5627,
20
- "step": 5000
21
- },
22
- {
23
- "epoch": 0.43,
24
- "eval_loss": 2.5240488052368164,
25
- "eval_runtime": 484.0815,
26
- "eval_samples_per_second": 9.519,
27
- "eval_steps_per_second": 4.76,
28
- "step": 5000
29
- },
30
- {
31
- "epoch": 0.64,
32
- "learning_rate": 7e-05,
33
- "loss": 2.4428,
34
- "step": 7500
35
- },
36
- {
37
- "epoch": 0.85,
38
- "learning_rate": 6e-05,
39
- "loss": 2.3709,
40
- "step": 10000
41
- },
42
- {
43
- "epoch": 0.85,
44
- "eval_loss": 2.3797976970672607,
45
- "eval_runtime": 484.2796,
46
- "eval_samples_per_second": 9.515,
47
- "eval_steps_per_second": 4.758,
48
- "step": 10000
49
- },
50
- {
51
- "epoch": 1.06,
52
- "learning_rate": 5e-05,
53
- "loss": 2.3225,
54
- "step": 12500
55
- },
56
- {
57
- "epoch": 1.28,
58
- "learning_rate": 4e-05,
59
- "loss": 2.2726,
60
- "step": 15000
61
- },
62
- {
63
- "epoch": 1.28,
64
- "eval_loss": 2.3043975830078125,
65
- "eval_runtime": 483.9345,
66
- "eval_samples_per_second": 9.522,
67
- "eval_steps_per_second": 4.761,
68
- "step": 15000
69
- },
70
- {
71
- "epoch": 1.49,
72
- "learning_rate": 3e-05,
73
- "loss": 2.2479,
74
- "step": 17500
75
- },
76
- {
77
- "epoch": 1.7,
78
- "learning_rate": 2e-05,
79
- "loss": 2.2341,
80
- "step": 20000
81
- },
82
- {
83
- "epoch": 1.7,
84
- "eval_loss": 2.264699935913086,
85
- "eval_runtime": 483.6102,
86
- "eval_samples_per_second": 9.528,
87
- "eval_steps_per_second": 4.764,
88
- "step": 20000
89
- },
90
- {
91
- "epoch": 1.91,
92
- "learning_rate": 1e-05,
93
- "loss": 2.2163,
94
- "step": 22500
95
- },
96
- {
97
- "epoch": 2.13,
98
- "learning_rate": 0.0,
99
- "loss": 2.2019,
100
- "step": 25000
101
- },
102
- {
103
- "epoch": 2.13,
104
- "eval_loss": 2.2437143325805664,
105
- "eval_runtime": 482.513,
106
- "eval_samples_per_second": 9.55,
107
- "eval_steps_per_second": 4.775,
108
- "step": 25000
109
- },
110
- {
111
- "epoch": 2.13,
112
- "step": 25000,
113
- "total_flos": 3.714827943936e+17,
114
- "train_loss": 2.3864154296875,
115
- "train_runtime": 60699.5161,
116
- "train_samples_per_second": 3.295,
117
- "train_steps_per_second": 0.412
118
- }
119
- ],
120
- "max_steps": 25000,
121
- "num_train_epochs": 3,
122
- "total_flos": 3.714827943936e+17,
123
- "trial_name": null,
124
- "trial_params": null
125
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:6581263b77f2005bb58dd0d702f7326a24b98a7257c0931e421b8398bc0be4b1
3
- size 3375
 
 
 
 
wikiann-az-results.txt DELETED
@@ -1,8 +0,0 @@
1
- ==================================================
2
- Results WikiANN-az
3
- ==================================================
4
- Model: /users/zyong2/data/zyong2/bigscience/data/processed/024/bloom-350m_az_fish_100000samples_-1vocab_original-frozen/
5
- [0.11019153464175929, 0.15776699029126215, 0.08903508771929824, 0.11444492070295757, 0.1271064034665383]
6
- 11.97
7
- 2.26
8
- ==================================================
 
 
 
 
 
 
 
 
 
word_embeddings.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:174ed618237771e5906be0e8d70c568de63633f3bb5e8a1e303bbdbaeaedc1ca
3
- size 1027605867
 
 
 
 
word_embeddings_layernorm.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:10917f86841a4f322406bd72ba2e4ae8e4780aaf462c98a76eca01e0c5fbc893
3
- size 9703