davidadamczyk commited on
Commit
d7399cb
·
verified ·
1 Parent(s): e04a55e

Model natrénován s nejlepší metrikou -0.0609 v epoše 0

Browse files
config.json ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "./output_test_freeze_dense_only/best_model",
3
+ "activation_function": "gelu",
4
+ "add_pooling_layer": false,
5
+ "architectures": [
6
+ "NomicBertModel"
7
+ ],
8
+ "attn_pdrop": 0.0,
9
+ "auto_map": {
10
+ "AutoConfig": "configuration_hf_nomic_bert.NomicBertConfig",
11
+ "AutoModel": "nomic-ai/nomic-bert-2048--modeling_hf_nomic_bert.NomicBertModel",
12
+ "AutoModelForMaskedLM": "nomic-ai/nomic-bert-2048--modeling_hf_nomic_bert.NomicBertForPreTraining",
13
+ "AutoModelForMultipleChoice": "nomic-ai/nomic-bert-2048--modeling_hf_nomic_bert.NomicBertForMultipleChoice",
14
+ "AutoModelForQuestionAnswering": "nomic-ai/nomic-bert-2048--modeling_hf_nomic_bert.NomicBertForQuestionAnswering",
15
+ "AutoModelForSequenceClassification": "nomic-ai/nomic-bert-2048--modeling_hf_nomic_bert.NomicBertForSequenceClassification",
16
+ "AutoModelForTokenClassification": "nomic-ai/nomic-bert-2048--modeling_hf_nomic_bert.NomicBertForTokenClassification"
17
+ },
18
+ "base_model_name": "./output_test_freeze_dense_only/best_model",
19
+ "bos_token_id": null,
20
+ "causal": false,
21
+ "dense_seq_output": true,
22
+ "embd_pdrop": 0.1,
23
+ "embedding_dim": 2048,
24
+ "eos_token_id": null,
25
+ "expert_choice_router": false,
26
+ "ffn_div": 1,
27
+ "freeze_embeddings": false,
28
+ "freeze_strategy": "selective",
29
+ "fused_bias_fc": true,
30
+ "fused_dropout_add_ln": true,
31
+ "initializer_range": 0.02,
32
+ "layer_norm_epsilon": 1e-05,
33
+ "max_trained_positions": 2048,
34
+ "mlp_fc1_bias": true,
35
+ "mlp_fc2_bias": true,
36
+ "model_type": "nomic_bert",
37
+ "moe_every_n_layers": 2,
38
+ "moe_impl": "megablocks",
39
+ "moe_normalize_expert_weights": false,
40
+ "moe_resid_pdrop": 0.0,
41
+ "moe_top_k": 2,
42
+ "n_embd": 768,
43
+ "n_head": 12,
44
+ "n_inner": 3072,
45
+ "n_layer": 12,
46
+ "n_positions": 2048,
47
+ "num_experts": 8,
48
+ "num_shared_experts": 0,
49
+ "num_unfrozen_layers": 6,
50
+ "pad_token_id": 1,
51
+ "pad_vocab_size_multiple": 64,
52
+ "parallel_block": false,
53
+ "parallel_block_tied_norm": false,
54
+ "prenorm": false,
55
+ "qkv_proj_bias": true,
56
+ "reorder_and_upcast_attn": false,
57
+ "resid_pdrop": 0.0,
58
+ "rotary_emb_base": 10000,
59
+ "rotary_emb_fraction": 1.0,
60
+ "rotary_emb_interleaved": false,
61
+ "rotary_emb_scale_base": null,
62
+ "rotary_scaling_factor": null,
63
+ "router_aux_loss_coef": 0.1,
64
+ "scale_attn_by_inverse_layer_idx": false,
65
+ "scale_attn_weights": true,
66
+ "summary_activation": null,
67
+ "summary_first_dropout": 0.1,
68
+ "summary_proj_to_labels": true,
69
+ "summary_type": "cls_index",
70
+ "summary_use_proj": true,
71
+ "torch_dtype": "float32",
72
+ "transformers_version": "4.49.0",
73
+ "type_vocab_size": 1,
74
+ "unfreeze_experts": false,
75
+ "use_cache": true,
76
+ "use_flash_attn": true,
77
+ "use_rms_norm": null,
78
+ "use_xentropy": true,
79
+ "vocab_size": 250048
80
+ }
configuration_hf_nomic_bert.py ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from transformers import GPT2Config
2
+
3
+
4
+ class NomicBertConfig(GPT2Config):
5
+ model_type = "nomic_bert"
6
+
7
+ def __init__(
8
+ self,
9
+ prenorm=False,
10
+ parallel_block=False,
11
+ parallel_block_tied_norm=False,
12
+ rotary_emb_fraction=0.0,
13
+ fused_dropout_add_ln=False,
14
+ fused_bias_fc=False,
15
+ use_flash_attn=False,
16
+ use_xentropy=False,
17
+ qkv_proj_bias=True,
18
+ rotary_emb_base=10_000,
19
+ rotary_emb_scale_base=None,
20
+ rotary_emb_interleaved=False,
21
+ mlp_fc1_bias=True,
22
+ mlp_fc2_bias=True,
23
+ use_rms_norm=False,
24
+ causal=False,
25
+ type_vocab_size=2,
26
+ dense_seq_output=True,
27
+ pad_vocab_size_multiple=1,
28
+ tie_word_embeddings=True,
29
+ rotary_scaling_factor=None,
30
+ max_trained_positions=2048,
31
+ **kwargs,
32
+ ):
33
+ self.prenorm = prenorm
34
+ self.parallel_block = parallel_block
35
+ self.parallel_block_tied_norm = parallel_block_tied_norm
36
+ self.rotary_emb_fraction = rotary_emb_fraction
37
+ self.tie_word_embeddings = tie_word_embeddings
38
+ self.fused_dropout_add_ln = fused_dropout_add_ln
39
+ self.fused_bias_fc = fused_bias_fc
40
+ self.use_flash_attn = use_flash_attn
41
+ self.use_xentropy = use_xentropy
42
+ self.qkv_proj_bias = qkv_proj_bias
43
+ self.rotary_emb_base = rotary_emb_base
44
+ self.rotary_emb_scale_base = rotary_emb_scale_base
45
+ self.rotary_emb_interleaved = rotary_emb_interleaved
46
+ self.mlp_fc1_bias = mlp_fc1_bias
47
+ self.mlp_fc2_bias = mlp_fc2_bias
48
+ self.use_rms_norm = use_rms_norm
49
+ self.causal = causal
50
+ self.type_vocab_size = type_vocab_size
51
+ self.dense_seq_output = dense_seq_output
52
+ self.pad_vocab_size_multiple = pad_vocab_size_multiple
53
+ self.rotary_scaling_factor = rotary_scaling_factor
54
+ self.max_trained_positions = max_trained_positions
55
+
56
+ super().__init__(**kwargs)
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc09ba7e0974054dfd397f3ec4028ec1793e3ed396e2e9131b31b0f2896cec31
3
+ size 1901187232
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01eb629a1c6c6166faadae1cbbd6f98224d172feab00fae51867d5efd1a57ce6
3
+ size 6318080