Synchronizing local compiler cache.
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/058e3ead348125e7808b.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/6bf9ef24cedd2d181630.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/7a4c62323b876a6ad082.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/gpt2/gpt2/780a42b2f8414c10d661.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/3033d959070fc853146f.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/46dc72bec418e48dc0b2.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/fdfe8a25cf9f41629cb9.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/llama/llamafactory/tiny-random-Llama-3/2e8cb42c92e334dfbaa9.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/llama/llamafactory/tiny-random-Llama-3/5433fffb45752137c8de.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/llama/llamafactory/tiny-random-Llama-3/8c48d1da1325809763e8.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/3afe2ef31d28f1e456be.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/aa9d07318c2ee53a79c9.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/ef87caaa517e593a73ba.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/mixtral/dacorvo/Mixtral-tiny/44ccd5195da49c90a57d.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/mixtral/dacorvo/Mixtral-tiny/908d5d307e281ceae80e.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/mixtral/dacorvo/Mixtral-tiny/be71598b4b4a5244a6a6.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/28192c462305efbf7703.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/62a1679ebaf9c75228d1.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/7e39204d4342f7703417.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/phi3/yujiepan/phi-4-tiny-random/33891c5b973ec2bfb226.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/phi3/yujiepan/phi-4-tiny-random/48d7f1c1e412ea84b60e.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/phi3/yujiepan/phi-4-tiny-random/a37d77e276a4a0f5c4e8.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/MODULE_032f74178031f5ed3c74+613edded/model.neff +1 -1
- neuronxcc-2.16.372.0+4a9b2326/MODULE_0a88901e8c98f54e4c10+613edded/model.neff +1 -1
- neuronxcc-2.16.372.0+4a9b2326/MODULE_0cfdfd8c26d66b282d5a+613edded/model.neff +0 -0
- neuronxcc-2.16.372.0+4a9b2326/MODULE_113f3268e3fd4d66fe81+8a3305d3/model.neff +0 -0
- neuronxcc-2.16.372.0+4a9b2326/MODULE_151e013069d6b102df91+613edded/model.neff +1 -1
- neuronxcc-2.16.372.0+4a9b2326/MODULE_24ff9ac2787ce9a1d276+613edded/model.neff +1 -1
- neuronxcc-2.16.372.0+4a9b2326/MODULE_272d7dabaeb48e6d7210+613edded/model.neff +1 -1
- neuronxcc-2.16.372.0+4a9b2326/MODULE_2b6914194b931d7496fc+613edded/model.neff +1 -1
- neuronxcc-2.16.372.0+4a9b2326/MODULE_2c290567bb51ca0069bf+613edded/model.neff +0 -0
- neuronxcc-2.16.372.0+4a9b2326/MODULE_320f2622d4d0c9fdd0f1+613edded/model.neff +1 -1
- neuronxcc-2.16.372.0+4a9b2326/MODULE_3d7f2d2bef4f6fdd2c74+613edded/model.neff +1 -1
- neuronxcc-2.16.372.0+4a9b2326/MODULE_3e5f6b34247d2b457ec5+613edded/model.neff +1 -1
- neuronxcc-2.16.372.0+4a9b2326/MODULE_3f0110aa8aef5f42c4bc+613edded/model.neff +1 -1
- neuronxcc-2.16.372.0+4a9b2326/MODULE_40f92bf9469aae653e93+613edded/model.neff +1 -1
- neuronxcc-2.16.372.0+4a9b2326/MODULE_48437cc74469a8ccaec8+613edded/model.neff +1 -1
- neuronxcc-2.16.372.0+4a9b2326/MODULE_506e29dd5cc46918936d+613edded/model.neff +1 -1
- neuronxcc-2.16.372.0+4a9b2326/MODULE_5de5ea910315ba9c0def+613edded/model.neff +1 -1
- neuronxcc-2.16.372.0+4a9b2326/MODULE_6819e64fb96e87ffece0+613edded/model.neff +1 -1
- neuronxcc-2.16.372.0+4a9b2326/MODULE_7062a76356ca462bcc78+613edded/model.neff +1 -1
- neuronxcc-2.16.372.0+4a9b2326/MODULE_71a034dd7c4a3afb59c6+613edded/model.neff +1 -1
- neuronxcc-2.16.372.0+4a9b2326/MODULE_7329da261de607372f14+613edded/model.neff +1 -1
- neuronxcc-2.16.372.0+4a9b2326/MODULE_75bc57af47ebdc36a75e+613edded/model.neff +1 -1
- neuronxcc-2.16.372.0+4a9b2326/MODULE_878400bd48be9c48dd23+613edded/model.neff +0 -0
- neuronxcc-2.16.372.0+4a9b2326/MODULE_8b38e41cf0a3c0152b87+613edded/model.neff +1 -1
- neuronxcc-2.16.372.0+4a9b2326/MODULE_8bb2b91314df12f54a63+613edded/model.neff +1 -1
- neuronxcc-2.16.372.0+4a9b2326/MODULE_960d05b842008ddeae63+8a3305d3/model.neff +0 -0
- neuronxcc-2.16.372.0+4a9b2326/MODULE_a297084c31f9c1ff1d5d+613edded/model.neff +1 -1
- neuronxcc-2.16.372.0+4a9b2326/MODULE_abb19eb6336ab05b7e19+613edded/model.neff +1 -1
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/058e3ead348125e7808b.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/6bf9ef24cedd2d181630.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/7a4c62323b876a6ad082.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/gpt2/gpt2/780a42b2f8414c10d661.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "architectures": ["GPT2LMHeadModel"], "attn_pdrop": 0.1, "bos_token_id": 50256, "embd_pdrop": 0.1, "eos_token_id": 50256, "initializer_range": 0.02, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 1024, "n_embd": 768, "n_head": 12, "n_inner": null, "n_layer": 12, "n_positions": 1024, "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "gpt2", "checkpoint_revision": "607a30d783dfa663caf39e06633721c8d4cfcd7e", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 1024, "task": "text-generation"}, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "task_specific_params": {"text-generation": {"do_sample": true, "max_length": 50}}, "use_cache": true, "vocab_size": 50257}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/3033d959070fc853146f.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/46dc72bec418e48dc0b2.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/fdfe8a25cf9f41629cb9.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/llama/llamafactory/tiny-random-Llama-3/2e8cb42c92e334dfbaa9.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": [128001, 128008, 128009], "head_dim": 4, "hidden_act": "silu", "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 64, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "llamafactory/tiny-random-Llama-3", "checkpoint_revision": "bf2a2e3bf199ad2ee96f02a3c00246c608db22a8", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 4, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 8.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 128256}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/llama/llamafactory/tiny-random-Llama-3/5433fffb45752137c8de.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": [128001, 128008, 128009], "head_dim": 4, "hidden_act": "silu", "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 64, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "llamafactory/tiny-random-Llama-3", "checkpoint_revision": "bf2a2e3bf199ad2ee96f02a3c00246c608db22a8", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 4, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 8.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 128256}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/llama/llamafactory/tiny-random-Llama-3/8c48d1da1325809763e8.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": [128001, 128008, 128009], "head_dim": 4, "hidden_act": "silu", "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 64, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "llamafactory/tiny-random-Llama-3", "checkpoint_revision": "bf2a2e3bf199ad2ee96f02a3c00246c608db22a8", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 4, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 8.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 128256}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/3afe2ef31d28f1e456be.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MistralForCausalLM"], "attention_dropout": 0.0, "attention_probs_dropout_prob": 0.1, "bos_token_id": 1, "eos_token_id": 2, "head_dim": 8, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 32, "initializer_range": 0.02, "intermediate_size": 37, "is_decoder": true, "max_position_embeddings": 512, "model_type": "mistral", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 2, "pad_token_id": 0, "rms_norm_eps": 1e-06, "rope_theta": 10000.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/aa9d07318c2ee53a79c9.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MistralForCausalLM"], "attention_dropout": 0.0, "attention_probs_dropout_prob": 0.1, "bos_token_id": 1, "eos_token_id": 2, "head_dim": 8, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 32, "initializer_range": 0.02, "intermediate_size": 37, "is_decoder": true, "max_position_embeddings": 512, "model_type": "mistral", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 2, "pad_token_id": 0, "rms_norm_eps": 1e-06, "rope_theta": 10000.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/ef87caaa517e593a73ba.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MistralForCausalLM"], "attention_dropout": 0.0, "attention_probs_dropout_prob": 0.1, "bos_token_id": 1, "eos_token_id": 2, "head_dim": 8, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 32, "initializer_range": 0.02, "intermediate_size": 37, "is_decoder": true, "max_position_embeddings": 512, "model_type": "mistral", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 2, "pad_token_id": 0, "rms_norm_eps": 1e-06, "rope_theta": 10000.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/mixtral/dacorvo/Mixtral-tiny/44ccd5195da49c90a57d.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MixtralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "head_dim": 32, "hidden_act": "silu", "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 3584, "max_position_embeddings": 1024, "model_type": "mixtral", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "dacorvo/Mixtral-tiny", "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 32, "num_experts_per_tok": 2, "num_hidden_layers": 2, "num_key_value_heads": 8, "num_local_experts": 8, "output_router_logits": false, "rms_norm_eps": 1e-05, "rope_theta": 10000.0, "router_aux_loss_coef": 0.001, "router_jitter_noise": 0.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/mixtral/dacorvo/Mixtral-tiny/908d5d307e281ceae80e.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MixtralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "head_dim": 32, "hidden_act": "silu", "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 3584, "max_position_embeddings": 1024, "model_type": "mixtral", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "dacorvo/Mixtral-tiny", "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 32, "num_experts_per_tok": 2, "num_hidden_layers": 2, "num_key_value_heads": 8, "num_local_experts": 8, "output_router_logits": false, "rms_norm_eps": 1e-05, "rope_theta": 10000.0, "router_aux_loss_coef": 0.001, "router_jitter_noise": 0.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/mixtral/dacorvo/Mixtral-tiny/be71598b4b4a5244a6a6.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MixtralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "head_dim": 32, "hidden_act": "silu", "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 3584, "max_position_embeddings": 1024, "model_type": "mixtral", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "dacorvo/Mixtral-tiny", "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 32, "num_experts_per_tok": 2, "num_hidden_layers": 2, "num_key_value_heads": 8, "num_local_experts": 8, "output_router_logits": false, "rms_norm_eps": 1e-05, "rope_theta": 10000.0, "router_aux_loss_coef": 0.001, "router_jitter_noise": 0.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/28192c462305efbf7703.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"_remove_final_layer_norm": false, "activation_function": "relu", "architectures": ["OPTForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 2, "do_layer_norm_before": true, "dropout": 0.1, "embed_dim": 16, "enable_bias": true, "eos_token_id": 2, "ffn_dim": 4, "hidden_size": 16, "init_std": 0.02, "is_decoder": true, "layer_norm_elementwise_affine": true, "layerdrop": 0.0, "max_position_embeddings": 100, "model_type": "opt", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "3f3f2ee1e499cb7ad89b877068684fdc9d9513c3", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 5, "pad_token_id": 1, "torch_dtype": "float32", "use_cache": true, "vocab_size": 50265, "word_embed_proj_dim": 16}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/62a1679ebaf9c75228d1.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"_remove_final_layer_norm": false, "activation_function": "relu", "architectures": ["OPTForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 2, "do_layer_norm_before": true, "dropout": 0.1, "embed_dim": 16, "enable_bias": true, "eos_token_id": 2, "ffn_dim": 4, "hidden_size": 16, "init_std": 0.02, "is_decoder": true, "layer_norm_elementwise_affine": true, "layerdrop": 0.0, "max_position_embeddings": 100, "model_type": "opt", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "3f3f2ee1e499cb7ad89b877068684fdc9d9513c3", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 5, "pad_token_id": 1, "torch_dtype": "float32", "use_cache": true, "vocab_size": 50265, "word_embed_proj_dim": 16}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/7e39204d4342f7703417.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"_remove_final_layer_norm": false, "activation_function": "relu", "architectures": ["OPTForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 2, "do_layer_norm_before": true, "dropout": 0.1, "embed_dim": 16, "enable_bias": true, "eos_token_id": 2, "ffn_dim": 4, "hidden_size": 16, "init_std": 0.02, "is_decoder": true, "layer_norm_elementwise_affine": true, "layerdrop": 0.0, "max_position_embeddings": 100, "model_type": "opt", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "3f3f2ee1e499cb7ad89b877068684fdc9d9513c3", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 5, "pad_token_id": 1, "torch_dtype": "float32", "use_cache": true, "vocab_size": 50265, "word_embed_proj_dim": 16}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/phi3/yujiepan/phi-4-tiny-random/33891c5b973ec2bfb226.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["Phi3ForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "auto_map": {}, "bos_token_id": 100257, "embd_pdrop": 0.0, "eos_token_id": 100257, "hidden_act": "silu", "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 32, "max_position_embeddings": 16384, "model_type": "phi3", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "yujiepan/phi-4-tiny-random", "checkpoint_revision": "18a9a1168dc97ac6d128f811925670c275610f5a", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 2, "num_hidden_layers": 2, "num_key_value_heads": 1, "original_max_position_embeddings": 16384, "pad_token_id": 100257, "resid_pdrop": 0.0, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 250000, "sliding_window": null, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 100352}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/phi3/yujiepan/phi-4-tiny-random/48d7f1c1e412ea84b60e.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["Phi3ForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "auto_map": {}, "bos_token_id": 100257, "embd_pdrop": 0.0, "eos_token_id": 100257, "hidden_act": "silu", "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 32, "max_position_embeddings": 16384, "model_type": "phi3", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "yujiepan/phi-4-tiny-random", "checkpoint_revision": "18a9a1168dc97ac6d128f811925670c275610f5a", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 2, "num_hidden_layers": 2, "num_key_value_heads": 1, "original_max_position_embeddings": 16384, "pad_token_id": 100257, "resid_pdrop": 0.0, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 250000, "sliding_window": null, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 100352}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.2.0.dev0/inference/phi3/yujiepan/phi-4-tiny-random/a37d77e276a4a0f5c4e8.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["Phi3ForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "auto_map": {}, "bos_token_id": 100257, "embd_pdrop": 0.0, "eos_token_id": 100257, "hidden_act": "silu", "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 32, "max_position_embeddings": 16384, "model_type": "phi3", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "yujiepan/phi-4-tiny-random", "checkpoint_revision": "18a9a1168dc97ac6d128f811925670c275610f5a", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 2, "num_hidden_layers": 2, "num_key_value_heads": 1, "original_max_position_embeddings": 16384, "pad_token_id": 100257, "resid_pdrop": 0.0, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 250000, "sliding_window": null, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 100352}
|
neuronxcc-2.16.372.0+4a9b2326/MODULE_032f74178031f5ed3c74+613edded/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 379904
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a85bc88aae6303157b499a09533d1d5c0cec207dbb414612e42392dd2b2787b8
|
3 |
size 379904
|
neuronxcc-2.16.372.0+4a9b2326/MODULE_0a88901e8c98f54e4c10+613edded/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 451584
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:724afe689cfb0e6019c30c5e8f47d9ca857e9836aa62692954d9d255912995b4
|
3 |
size 451584
|
neuronxcc-2.16.372.0+4a9b2326/MODULE_0cfdfd8c26d66b282d5a+613edded/model.neff
CHANGED
Binary files a/neuronxcc-2.16.372.0+4a9b2326/MODULE_0cfdfd8c26d66b282d5a+613edded/model.neff and b/neuronxcc-2.16.372.0+4a9b2326/MODULE_0cfdfd8c26d66b282d5a+613edded/model.neff differ
|
|
neuronxcc-2.16.372.0+4a9b2326/MODULE_113f3268e3fd4d66fe81+8a3305d3/model.neff
CHANGED
Binary files a/neuronxcc-2.16.372.0+4a9b2326/MODULE_113f3268e3fd4d66fe81+8a3305d3/model.neff and b/neuronxcc-2.16.372.0+4a9b2326/MODULE_113f3268e3fd4d66fe81+8a3305d3/model.neff differ
|
|
neuronxcc-2.16.372.0+4a9b2326/MODULE_151e013069d6b102df91+613edded/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 123904
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cec99f24e396492581bfdca52ea60b5a7f3e8e34386abf0193d7963614f6992d
|
3 |
size 123904
|
neuronxcc-2.16.372.0+4a9b2326/MODULE_24ff9ac2787ce9a1d276+613edded/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 154624
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9c9efa8b2c66754b5bcd292b6ef26abb1e46f524440fd49d403b0ec92ee90bff
|
3 |
size 154624
|
neuronxcc-2.16.372.0+4a9b2326/MODULE_272d7dabaeb48e6d7210+613edded/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 123904
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:82b76d859c8147cd419abf81056f2959376de57ca3b683d212a6346dddf161c6
|
3 |
size 123904
|
neuronxcc-2.16.372.0+4a9b2326/MODULE_2b6914194b931d7496fc+613edded/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 410624
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0906e8bf96be2119a2e1265dcafee4b8f061eac376aedaba6d7bd00ff4ba9358
|
3 |
size 410624
|
neuronxcc-2.16.372.0+4a9b2326/MODULE_2c290567bb51ca0069bf+613edded/model.neff
CHANGED
Binary files a/neuronxcc-2.16.372.0+4a9b2326/MODULE_2c290567bb51ca0069bf+613edded/model.neff and b/neuronxcc-2.16.372.0+4a9b2326/MODULE_2c290567bb51ca0069bf+613edded/model.neff differ
|
|
neuronxcc-2.16.372.0+4a9b2326/MODULE_320f2622d4d0c9fdd0f1+613edded/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 144384
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:715fba7c6f4f81d01d141d02d49351ce649988bb343f8eaf8d22b95620645c0e
|
3 |
size 144384
|
neuronxcc-2.16.372.0+4a9b2326/MODULE_3d7f2d2bef4f6fdd2c74+613edded/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 154624
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6fe88b03e5ae3c0243d27bdcefaa05ea7c17a8aa36d4c8da4e33664f1fb14bc4
|
3 |
size 154624
|
neuronxcc-2.16.372.0+4a9b2326/MODULE_3e5f6b34247d2b457ec5+613edded/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 134144
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:05c4b96f997ac0320a92a2e16b5e39dc55c18c696524cc8e0522e6846139c1e0
|
3 |
size 134144
|
neuronxcc-2.16.372.0+4a9b2326/MODULE_3f0110aa8aef5f42c4bc+613edded/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 144384
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d41e8505bcaa3d37c93b97df2961c96c40a2c6d88afff9e9acad2af30433367f
|
3 |
size 144384
|
neuronxcc-2.16.372.0+4a9b2326/MODULE_40f92bf9469aae653e93+613edded/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 113664
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2b7b417596af9f82a293ae44584e98965ca5e0ae85d2d1894accfb76bdd00696
|
3 |
size 113664
|
neuronxcc-2.16.372.0+4a9b2326/MODULE_48437cc74469a8ccaec8+613edded/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 154624
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8cd10b9bb9f25defa96fbc21ad0a1b79c2b4164389960b8eaae2c70621d2e278
|
3 |
size 154624
|
neuronxcc-2.16.372.0+4a9b2326/MODULE_506e29dd5cc46918936d+613edded/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 134144
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a0e17054c1a018904931a4f4ec4dba8678877b4578e837919c3705c4df3ca211
|
3 |
size 134144
|
neuronxcc-2.16.372.0+4a9b2326/MODULE_5de5ea910315ba9c0def+613edded/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 134144
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b1cd7c449df226b9275fe94e3b749ec6c1bae24dc8ee6ba9013b35ec6548b97d
|
3 |
size 134144
|
neuronxcc-2.16.372.0+4a9b2326/MODULE_6819e64fb96e87ffece0+613edded/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 144384
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:be4a07f7917df7a74381ee98a03c313b3cb17c4e25dfa9657654157bcc732aef
|
3 |
size 144384
|
neuronxcc-2.16.372.0+4a9b2326/MODULE_7062a76356ca462bcc78+613edded/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 134144
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0d8b2164c08732f68c06ea9207ce61a94e80bf7dca0896ebc1c335bd27e38f76
|
3 |
size 134144
|
neuronxcc-2.16.372.0+4a9b2326/MODULE_71a034dd7c4a3afb59c6+613edded/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 123904
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e4c9ab2f575f45527965a9d26cd88725b032408246ec533cf7e995d136d7d520
|
3 |
size 123904
|
neuronxcc-2.16.372.0+4a9b2326/MODULE_7329da261de607372f14+613edded/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 134144
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f3395c25942505761ef6f0379f37023044366f80e9c0e10a4bfd97eb11888894
|
3 |
size 134144
|
neuronxcc-2.16.372.0+4a9b2326/MODULE_75bc57af47ebdc36a75e+613edded/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 123904
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9c8befd4c5d7a093ed670384eadadd239921e964407ef85f896ef906b73c5de3
|
3 |
size 123904
|
neuronxcc-2.16.372.0+4a9b2326/MODULE_878400bd48be9c48dd23+613edded/model.neff
CHANGED
Binary files a/neuronxcc-2.16.372.0+4a9b2326/MODULE_878400bd48be9c48dd23+613edded/model.neff and b/neuronxcc-2.16.372.0+4a9b2326/MODULE_878400bd48be9c48dd23+613edded/model.neff differ
|
|
neuronxcc-2.16.372.0+4a9b2326/MODULE_8b38e41cf0a3c0152b87+613edded/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 123904
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b74554f57fc9c918cd61d432e7b4c1612ca7288bad35a1d3ff3cd84907f41a1b
|
3 |
size 123904
|
neuronxcc-2.16.372.0+4a9b2326/MODULE_8bb2b91314df12f54a63+613edded/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 410624
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1efbaef8da21acffa89375cbaf8ce791f1d64a2c49cbf59e5c2e3c75ce0c901f
|
3 |
size 410624
|
neuronxcc-2.16.372.0+4a9b2326/MODULE_960d05b842008ddeae63+8a3305d3/model.neff
CHANGED
Binary files a/neuronxcc-2.16.372.0+4a9b2326/MODULE_960d05b842008ddeae63+8a3305d3/model.neff and b/neuronxcc-2.16.372.0+4a9b2326/MODULE_960d05b842008ddeae63+8a3305d3/model.neff differ
|
|
neuronxcc-2.16.372.0+4a9b2326/MODULE_a297084c31f9c1ff1d5d+613edded/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 154624
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:29e1796f5c44e8f24a9c9075037799af89efba73cea49227c437284b0fc7ebca
|
3 |
size 154624
|
neuronxcc-2.16.372.0+4a9b2326/MODULE_abb19eb6336ab05b7e19+613edded/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 410624
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6c87f066c6f997672764a455beb6aa2e888c97c05c9f253026eb91a1ccbffd2a
|
3 |
size 410624
|