Synchronizing local compiler cache.
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +81 -0
- neuronxcc-2.15.143.0+e39249ad/MODULE_01e529942e52a9f61be4+39f12043/compile_flags.json +1 -0
- neuronxcc-2.15.143.0+e39249ad/MODULE_01e529942e52a9f61be4+39f12043/model.done +0 -0
- neuronxcc-2.15.143.0+e39249ad/MODULE_01e529942e52a9f61be4+39f12043/model.hlo_module.pb +3 -0
- neuronxcc-2.15.143.0+e39249ad/MODULE_01e529942e52a9f61be4+39f12043/model.neff +3 -0
- neuronxcc-2.15.143.0+e39249ad/MODULE_9a2a6520ed76b8de514b+39f12043/compile_flags.json +1 -0
- neuronxcc-2.15.143.0+e39249ad/MODULE_9a2a6520ed76b8de514b+39f12043/model.done +0 -0
- neuronxcc-2.15.143.0+e39249ad/MODULE_9a2a6520ed76b8de514b+39f12043/model.hlo_module.pb +3 -0
- neuronxcc-2.15.143.0+e39249ad/MODULE_9a2a6520ed76b8de514b+39f12043/model.neff +3 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/058e3ead348125e7808b.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/6bf9ef24cedd2d181630.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/7a4c62323b876a6ad082.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/gpt2/gpt2/780a42b2f8414c10d661.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/gpt2/hf-internal-testing/tiny-random-gpt2/3033d959070fc853146f.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/gpt2/hf-internal-testing/tiny-random-gpt2/46dc72bec418e48dc0b2.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/gpt2/hf-internal-testing/tiny-random-gpt2/fdfe8a25cf9f41629cb9.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/granite/ibm-granite/granite-3.1-2b-instruct/bf71032506a251068193.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/llama/Maykeye/TinyLLama-v0/f71b7fdfb0dfed1fac13.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/llama/NousResearch/Hermes-2-Theta-Llama-3-8B/7bae97d51948b959db6b.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/llama/llamafactory/tiny-random-Llama-3/2e8cb42c92e334dfbaa9.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/llama/llamafactory/tiny-random-Llama-3/5433fffb45752137c8de.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/llama/llamafactory/tiny-random-Llama-3/8c48d1da1325809763e8.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/llama/unsloth/Llama-3.2-1B-Instruct/49d61e18927b113552d1.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/3afe2ef31d28f1e456be.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/aa9d07318c2ee53a79c9.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/ef87caaa517e593a73ba.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/mistral/optimum/mistral-1.1b-testing/6668357e5d6bb2f40fe7.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/mixtral/dacorvo/Mixtral-tiny/44ccd5195da49c90a57d.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/mixtral/dacorvo/Mixtral-tiny/908d5d307e281ceae80e.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/mixtral/dacorvo/Mixtral-tiny/be71598b4b4a5244a6a6.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/mixtral/dacorvo/Mixtral-tiny/f574775a377747e5836b.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/28192c462305efbf7703.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/62a1679ebaf9c75228d1.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/7e39204d4342f7703417.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/phi3/microsoft/Phi-3-mini-4k-instruct/34e681f1c4c2cc6c581b.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/phi3/microsoft/Phi-3-mini-4k-instruct/91032bbebe8c505cd37f.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/phi3/microsoft/Phi-3-mini-4k-instruct/d7892042a1da5eaa0020.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/phi3/microsoft/Phi-3-mini-4k-instruct/f6d3183f58f8544aad0f.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/phi3/yujiepan/phi-4-tiny-random/33891c5b973ec2bfb226.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/phi3/yujiepan/phi-4-tiny-random/48d7f1c1e412ea84b60e.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/phi3/yujiepan/phi-4-tiny-random/a37d77e276a4a0f5c4e8.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/qwen2/Qwen/Qwen2.5-0.5B/08b82ea5dab18178ad7f.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev3/inference/gpt2/gpt2/780a42b2f8414c10d661.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev3/inference/granite/ibm-granite/granite-3.1-2b-instruct/bf71032506a251068193.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev3/inference/llama/meta-llama/Llama-3.2-1B-Instruct/c19d11f088aafb7b1382.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev3/inference/llama/unsloth/Llama-3.2-1B-Instruct/49d61e18927b113552d1.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev3/inference/llama/unsloth/Llama-3.2-1B-Instruct/9d4dd76cfd7e72235824.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev3/inference/llama/unsloth/Llama-3.2-3B-Instruct/b6352a1b005c1b4fdbee.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev3/inference/mistral/optimum/mistral-1.1b-testing/6668357e5d6bb2f40fe7.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev3/inference/mixtral/dacorvo/Mixtral-tiny/f574775a377747e5836b.json +1 -0
.gitattributes
CHANGED
@@ -7504,3 +7504,84 @@ neuronxcc-2.16.372.0+4a9b2326/MODULE_94df5568a6ee6b8f721b+613edded/model.neff fi
|
|
7504 |
neuronxcc-2.16.372.0+4a9b2326/MODULE_9dc8af288609e8374339+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7505 |
neuronxcc-2.16.372.0+4a9b2326/MODULE_a9bcc20735f75ac1f46e+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7506 |
neuronxcc-2.16.372.0+4a9b2326/MODULE_b2821092d30cdeb76d6a+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
7504 |
neuronxcc-2.16.372.0+4a9b2326/MODULE_9dc8af288609e8374339+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7505 |
neuronxcc-2.16.372.0+4a9b2326/MODULE_a9bcc20735f75ac1f46e+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7506 |
neuronxcc-2.16.372.0+4a9b2326/MODULE_b2821092d30cdeb76d6a+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7507 |
+
neuronxcc-2.15.143.0+e39249ad/MODULE_01e529942e52a9f61be4+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
7508 |
+
neuronxcc-2.15.143.0+e39249ad/MODULE_9a2a6520ed76b8de514b+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
7509 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_00804590004fa5eab7c9+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7510 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_00c7ea8a2ca42389a896+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7511 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_0295f5e523e3ea7df756+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7512 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_032f74178031f5ed3c74+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7513 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_05b0003a35326e3b2f57+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
7514 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_0977746a68f18f7b9228+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7515 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_09d665c9864fd54d66b2+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7516 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_0a88901e8c98f54e4c10+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7517 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_0c3610ac3206c8024aae+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7518 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_151e013069d6b102df91+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7519 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_177e7d4bfa5feb0f544b+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7520 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_19afba81b3e0cb8f888e+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
7521 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_20f9276a7330dc649731+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
7522 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_24ff9ac2787ce9a1d276+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7523 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_25c42f91982701aa03cc+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
7524 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_272d7dabaeb48e6d7210+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7525 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_2b6914194b931d7496fc+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7526 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_2d763f4e10d60c4030bd+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
7527 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_2db40b5b23a523e16b87+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7528 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_320f2622d4d0c9fdd0f1+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7529 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_396f866b25e26394047c+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7530 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_3a070e0dd0e8091b4ebd+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
7531 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_3d7f2d2bef4f6fdd2c74+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7532 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_3e5f6b34247d2b457ec5+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7533 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_3f0110aa8aef5f42c4bc+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7534 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_40f92bf9469aae653e93+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7535 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_48437cc74469a8ccaec8+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7536 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_4cccf279275b39cca49c+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7537 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_4e6317a710ff2fcf60cd+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7538 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_506e29dd5cc46918936d+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7539 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_531613db5d175a66b951+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7540 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_5371a46c9607cc9aeba7+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7541 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_5b36e3b39e7f0fc4612b+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7542 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_5de5ea910315ba9c0def+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7543 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_5e35dbcf8ed725f26ec0+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
7544 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_6819e64fb96e87ffece0+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7545 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_682689dd0dabdac18fbd+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7546 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_686d3373c8eb03797f41+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7547 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_7062a76356ca462bcc78+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7548 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_71a034dd7c4a3afb59c6+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7549 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_7329da261de607372f14+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7550 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_73e7362a3d706e4803fc+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
7551 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_7418d879b0b0dbe9d053+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7552 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_75bc57af47ebdc36a75e+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7553 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_7ce64f3fbb3960eec319+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
7554 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_7fce147fef3e3e4e1386+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7555 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_80ed60eb29ade17d45ff+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7556 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_897fc8ee0fee573d4294+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7557 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_8a17d7b25a214e612693+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
7558 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_8b38e41cf0a3c0152b87+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7559 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_8bb2b91314df12f54a63+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7560 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_922ce3591b342750be6b+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7561 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_934e6cd7e79b2fea88ab+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7562 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_96624f9b555d2ee1123d+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7563 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_967d83f73b8fe30e4d14+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7564 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_9a2a6520ed76b8de514b+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
7565 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_9d4538dd1cfb96628706+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7566 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_a08a6102a96eea257396+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7567 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_a297084c31f9c1ff1d5d+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7568 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_a31729227d8ec6b4f136+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7569 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_a95326667095c4d75db0+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7570 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_abb19eb6336ab05b7e19+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7571 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_ad7506c3fc27d75f26a6+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
7572 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_be5892d16c7e2fb6594c+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
7573 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_bf70065b1150aaeca3bc+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7574 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_c1ff08289224c1071721+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7575 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_c49c0c3715f68c22b32f+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7576 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_cc8cf237a3eedbaaeb9b+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7577 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_dcc643bd43691cdb2cd2+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7578 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_dd4a13f50d0dab4871b8+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7579 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_de5fab2e2ecc60551050+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7580 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_df48af4bf01af7f3857e+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7581 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_e50c30ad9376ce18d039+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7582 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_e8eb6214f5c387ad6b43+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7583 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_e99dd53d1f163d6bb260+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7584 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_eaebd6f723294af70148+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7585 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_f7bd275f6b204e1d6808+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
7586 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_f91cd98a64f373af274d+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
7587 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_feed1320e32bafc2a577+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
neuronxcc-2.15.143.0+e39249ad/MODULE_01e529942e52a9f61be4+39f12043/compile_flags.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
["--model-type=transformer", "--auto-cast=none", "--execute-repetition=1"]
|
neuronxcc-2.15.143.0+e39249ad/MODULE_01e529942e52a9f61be4+39f12043/model.done
ADDED
File without changes
|
neuronxcc-2.15.143.0+e39249ad/MODULE_01e529942e52a9f61be4+39f12043/model.hlo_module.pb
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e515f0718c75625a19063d4f578f532de942a2d35948a292fc17ac27c7e0c91c
|
3 |
+
size 430651
|
neuronxcc-2.15.143.0+e39249ad/MODULE_01e529942e52a9f61be4+39f12043/model.neff
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:65982f2734cd853c06302c658436d23c755d95866603e54ccfe01315a664d7da
|
3 |
+
size 22866944
|
neuronxcc-2.15.143.0+e39249ad/MODULE_9a2a6520ed76b8de514b+39f12043/compile_flags.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
["--model-type=transformer", "--auto-cast=none", "--execute-repetition=1"]
|
neuronxcc-2.15.143.0+e39249ad/MODULE_9a2a6520ed76b8de514b+39f12043/model.done
ADDED
File without changes
|
neuronxcc-2.15.143.0+e39249ad/MODULE_9a2a6520ed76b8de514b+39f12043/model.hlo_module.pb
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b6f4316db50af657375dfaa4c122cafee102332b02b8527169bccfbb7df2eac6
|
3 |
+
size 338627
|
neuronxcc-2.15.143.0+e39249ad/MODULE_9a2a6520ed76b8de514b+39f12043/model.neff
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ca6d7e3b9b3b485a5ba7f4d68b4853d8060b52853725eb2b129271483b1d7ad3
|
3 |
+
size 2407424
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/058e3ead348125e7808b.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/6bf9ef24cedd2d181630.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/7a4c62323b876a6ad082.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/gpt2/gpt2/780a42b2f8414c10d661.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "architectures": ["GPT2LMHeadModel"], "attn_pdrop": 0.1, "bos_token_id": 50256, "embd_pdrop": 0.1, "eos_token_id": 50256, "initializer_range": 0.02, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 1024, "n_embd": 768, "n_head": 12, "n_inner": null, "n_layer": 12, "n_positions": 1024, "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "gpt2", "checkpoint_revision": "607a30d783dfa663caf39e06633721c8d4cfcd7e", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 1024, "task": "text-generation"}, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "task_specific_params": {"text-generation": {"do_sample": true, "max_length": 50}}, "use_cache": true, "vocab_size": 50257}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/gpt2/hf-internal-testing/tiny-random-gpt2/3033d959070fc853146f.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/gpt2/hf-internal-testing/tiny-random-gpt2/46dc72bec418e48dc0b2.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/gpt2/hf-internal-testing/tiny-random-gpt2/fdfe8a25cf9f41629cb9.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/granite/ibm-granite/granite-3.1-2b-instruct/bf71032506a251068193.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["GraniteForCausalLM"], "attention_bias": false, "attention_dropout": 0.1, "attention_multiplier": 0.015625, "bos_token_id": 0, "embedding_multiplier": 12.0, "eos_token_id": 0, "hidden_act": "silu", "hidden_size": 2048, "initializer_range": 0.02, "intermediate_size": 8192, "logits_scaling": 8.0, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "granite", "neuron": {"auto_cast_type": "bf16", "batch_size": 4, "checkpoint_id": "ibm-granite/granite-3.1-2b-instruct", "checkpoint_revision": "374ef54e020a3ce208c65e96d6213922a87d8952", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 40, "num_key_value_heads": 8, "pad_token_id": 0, "residual_multiplier": 0.22, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 5000000.0, "tie_word_embeddings": true, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 49155}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/llama/Maykeye/TinyLLama-v0/f71b7fdfb0dfed1fac13.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "head_dim": 4, "hidden_act": "silu", "hidden_size": 64, "initializer_range": 0.02, "intermediate_size": 256, "max_position_embeddings": 2048, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "bf16", "batch_size": 2, "checkpoint_id": "Maykeye/TinyLLama-v0", "checkpoint_revision": "298338802ab94432b917bcce11382aa151aee50f", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 1024, "task": "text-generation"}, "num_attention_heads": 16, "num_hidden_layers": 8, "num_key_value_heads": 16, "pad_token_id": 0, "pretraining_tp": 1, "rms_norm_eps": 1e-06, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/llama/NousResearch/Hermes-2-Theta-Llama-3-8B/7bae97d51948b959db6b.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": 128003, "head_dim": 128, "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 14336, "max_position_embeddings": 8192, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "NousResearch/Hermes-2-Theta-Llama-3-8B", "checkpoint_revision": "57a73110702e7b05ba3f39fef36297454c680725", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 128256}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/llama/llamafactory/tiny-random-Llama-3/2e8cb42c92e334dfbaa9.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": [128001, 128008, 128009], "head_dim": 4, "hidden_act": "silu", "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 64, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "llamafactory/tiny-random-Llama-3", "checkpoint_revision": "bf2a2e3bf199ad2ee96f02a3c00246c608db22a8", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 4, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 8.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 128256}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/llama/llamafactory/tiny-random-Llama-3/5433fffb45752137c8de.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": [128001, 128008, 128009], "head_dim": 4, "hidden_act": "silu", "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 64, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "llamafactory/tiny-random-Llama-3", "checkpoint_revision": "bf2a2e3bf199ad2ee96f02a3c00246c608db22a8", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 4, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 8.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 128256}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/llama/llamafactory/tiny-random-Llama-3/8c48d1da1325809763e8.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": [128001, 128008, 128009], "head_dim": 4, "hidden_act": "silu", "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 64, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "llamafactory/tiny-random-Llama-3", "checkpoint_revision": "bf2a2e3bf199ad2ee96f02a3c00246c608db22a8", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 4, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 8.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 128256}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/llama/unsloth/Llama-3.2-1B-Instruct/49d61e18927b113552d1.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": 128009, "head_dim": 64, "hidden_act": "silu", "hidden_size": 2048, "initializer_range": 0.02, "intermediate_size": 8192, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "unsloth/Llama-3.2-1B-Instruct", "checkpoint_revision": "9b58d4a36161a1e49ecf0a69d20b2736fef8e438", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 16, "num_key_value_heads": 8, "pad_token_id": 128004, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 32.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": true, "torch_dtype": "bfloat16", "unsloth_fixed": true, "use_cache": true, "vocab_size": 128256}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/3afe2ef31d28f1e456be.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MistralForCausalLM"], "attention_dropout": 0.0, "attention_probs_dropout_prob": 0.1, "bos_token_id": 1, "eos_token_id": 2, "head_dim": 8, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 32, "initializer_range": 0.02, "intermediate_size": 37, "is_decoder": true, "max_position_embeddings": 512, "model_type": "mistral", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 2, "pad_token_id": 0, "rms_norm_eps": 1e-06, "rope_theta": 10000.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/aa9d07318c2ee53a79c9.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MistralForCausalLM"], "attention_dropout": 0.0, "attention_probs_dropout_prob": 0.1, "bos_token_id": 1, "eos_token_id": 2, "head_dim": 8, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 32, "initializer_range": 0.02, "intermediate_size": 37, "is_decoder": true, "max_position_embeddings": 512, "model_type": "mistral", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 2, "pad_token_id": 0, "rms_norm_eps": 1e-06, "rope_theta": 10000.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/ef87caaa517e593a73ba.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MistralForCausalLM"], "attention_dropout": 0.0, "attention_probs_dropout_prob": 0.1, "bos_token_id": 1, "eos_token_id": 2, "head_dim": 8, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 32, "initializer_range": 0.02, "intermediate_size": 37, "is_decoder": true, "max_position_embeddings": 512, "model_type": "mistral", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 2, "pad_token_id": 0, "rms_norm_eps": 1e-06, "rope_theta": 10000.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/mistral/optimum/mistral-1.1b-testing/6668357e5d6bb2f40fe7.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MistralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "head_dim": 64, "hidden_act": "silu", "hidden_size": 2048, "initializer_range": 0.02, "intermediate_size": 5632, "max_position_embeddings": 32768, "model_type": "mistral", "neuron": {"auto_cast_type": "bf16", "batch_size": 4, "checkpoint_id": "optimum/mistral-1.1b-testing", "checkpoint_revision": "ce03bc8d47dbd2c173ff65f3a8de1325ba724195", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 22, "num_key_value_heads": 4, "rms_norm_eps": 1e-05, "rope_theta": 1000000.0, "sliding_window": null, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/mixtral/dacorvo/Mixtral-tiny/44ccd5195da49c90a57d.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MixtralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "head_dim": 32, "hidden_act": "silu", "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 3584, "max_position_embeddings": 1024, "model_type": "mixtral", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "dacorvo/Mixtral-tiny", "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 32, "num_experts_per_tok": 2, "num_hidden_layers": 2, "num_key_value_heads": 8, "num_local_experts": 8, "output_router_logits": false, "rms_norm_eps": 1e-05, "rope_theta": 10000.0, "router_aux_loss_coef": 0.001, "router_jitter_noise": 0.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/mixtral/dacorvo/Mixtral-tiny/908d5d307e281ceae80e.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MixtralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "head_dim": 32, "hidden_act": "silu", "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 3584, "max_position_embeddings": 1024, "model_type": "mixtral", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "dacorvo/Mixtral-tiny", "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 32, "num_experts_per_tok": 2, "num_hidden_layers": 2, "num_key_value_heads": 8, "num_local_experts": 8, "output_router_logits": false, "rms_norm_eps": 1e-05, "rope_theta": 10000.0, "router_aux_loss_coef": 0.001, "router_jitter_noise": 0.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/mixtral/dacorvo/Mixtral-tiny/be71598b4b4a5244a6a6.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MixtralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "head_dim": 32, "hidden_act": "silu", "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 3584, "max_position_embeddings": 1024, "model_type": "mixtral", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "dacorvo/Mixtral-tiny", "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 32, "num_experts_per_tok": 2, "num_hidden_layers": 2, "num_key_value_heads": 8, "num_local_experts": 8, "output_router_logits": false, "rms_norm_eps": 1e-05, "rope_theta": 10000.0, "router_aux_loss_coef": 0.001, "router_jitter_noise": 0.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/mixtral/dacorvo/Mixtral-tiny/f574775a377747e5836b.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MixtralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "head_dim": 32, "hidden_act": "silu", "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 3584, "max_position_embeddings": 1024, "model_type": "mixtral", "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "dacorvo/Mixtral-tiny", "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 1024, "task": "text-generation"}, "num_attention_heads": 32, "num_experts_per_tok": 2, "num_hidden_layers": 2, "num_key_value_heads": 8, "num_local_experts": 8, "output_router_logits": false, "rms_norm_eps": 1e-05, "rope_theta": 10000.0, "router_aux_loss_coef": 0.001, "router_jitter_noise": 0.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/28192c462305efbf7703.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"_remove_final_layer_norm": false, "activation_function": "relu", "architectures": ["OPTForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 2, "do_layer_norm_before": true, "dropout": 0.1, "embed_dim": 16, "enable_bias": true, "eos_token_id": 2, "ffn_dim": 4, "hidden_size": 16, "init_std": 0.02, "is_decoder": true, "layer_norm_elementwise_affine": true, "layerdrop": 0.0, "max_position_embeddings": 100, "model_type": "opt", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "3f3f2ee1e499cb7ad89b877068684fdc9d9513c3", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 5, "pad_token_id": 1, "torch_dtype": "float32", "use_cache": true, "vocab_size": 50265, "word_embed_proj_dim": 16}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/62a1679ebaf9c75228d1.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"_remove_final_layer_norm": false, "activation_function": "relu", "architectures": ["OPTForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 2, "do_layer_norm_before": true, "dropout": 0.1, "embed_dim": 16, "enable_bias": true, "eos_token_id": 2, "ffn_dim": 4, "hidden_size": 16, "init_std": 0.02, "is_decoder": true, "layer_norm_elementwise_affine": true, "layerdrop": 0.0, "max_position_embeddings": 100, "model_type": "opt", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "3f3f2ee1e499cb7ad89b877068684fdc9d9513c3", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 5, "pad_token_id": 1, "torch_dtype": "float32", "use_cache": true, "vocab_size": 50265, "word_embed_proj_dim": 16}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/7e39204d4342f7703417.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"_remove_final_layer_norm": false, "activation_function": "relu", "architectures": ["OPTForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 2, "do_layer_norm_before": true, "dropout": 0.1, "embed_dim": 16, "enable_bias": true, "eos_token_id": 2, "ffn_dim": 4, "hidden_size": 16, "init_std": 0.02, "is_decoder": true, "layer_norm_elementwise_affine": true, "layerdrop": 0.0, "max_position_embeddings": 100, "model_type": "opt", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "3f3f2ee1e499cb7ad89b877068684fdc9d9513c3", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 5, "pad_token_id": 1, "torch_dtype": "float32", "use_cache": true, "vocab_size": 50265, "word_embed_proj_dim": 16}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/phi3/microsoft/Phi-3-mini-4k-instruct/34e681f1c4c2cc6c581b.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["Phi3ForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "auto_map": {"AutoConfig": "microsoft/Phi-3-mini-4k-instruct--configuration_phi3.Phi3Config", "AutoModelForCausalLM": "microsoft/Phi-3-mini-4k-instruct--modeling_phi3.Phi3ForCausalLM"}, "bos_token_id": 1, "embd_pdrop": 0.0, "eos_token_id": 32000, "hidden_act": "silu", "hidden_size": 3072, "initializer_range": 0.02, "intermediate_size": 8192, "max_position_embeddings": 4096, "model_type": "phi3", "neuron": {"auto_cast_type": "bf16", "batch_size": 4, "checkpoint_id": "microsoft/Phi-3-mini-4k-instruct", "checkpoint_revision": "0a67737cc96d2554230f90338b163bc6380a2a85", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 32, "original_max_position_embeddings": 4096, "pad_token_id": 32000, "resid_pdrop": 0.0, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000.0, "sliding_window": 2047, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 32064}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/phi3/microsoft/Phi-3-mini-4k-instruct/91032bbebe8c505cd37f.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["Phi3ForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "auto_map": {"AutoConfig": "microsoft/Phi-3-mini-4k-instruct--configuration_phi3.Phi3Config", "AutoModelForCausalLM": "microsoft/Phi-3-mini-4k-instruct--modeling_phi3.Phi3ForCausalLM"}, "bos_token_id": 1, "embd_pdrop": 0.0, "eos_token_id": 32000, "hidden_act": "silu", "hidden_size": 3072, "initializer_range": 0.02, "intermediate_size": 8192, "max_position_embeddings": 4096, "model_type": "phi3", "neuron": {"auto_cast_type": "bf16", "batch_size": 2, "checkpoint_id": "microsoft/Phi-3-mini-4k-instruct", "checkpoint_revision": "0a67737cc96d2554230f90338b163bc6380a2a85", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 4, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 32, "original_max_position_embeddings": 4096, "pad_token_id": 32000, "resid_pdrop": 0.0, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000.0, "sliding_window": 2047, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 32064}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/phi3/microsoft/Phi-3-mini-4k-instruct/d7892042a1da5eaa0020.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["Phi3ForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "auto_map": {"AutoConfig": "microsoft/Phi-3-mini-4k-instruct--configuration_phi3.Phi3Config", "AutoModelForCausalLM": "microsoft/Phi-3-mini-4k-instruct--modeling_phi3.Phi3ForCausalLM"}, "bos_token_id": 1, "embd_pdrop": 0.0, "eos_token_id": 32000, "hidden_act": "silu", "hidden_size": 3072, "initializer_range": 0.02, "intermediate_size": 8192, "max_position_embeddings": 4096, "model_type": "phi3", "neuron": {"auto_cast_type": "bf16", "batch_size": 2, "checkpoint_id": "microsoft/Phi-3-mini-4k-instruct", "checkpoint_revision": "0a67737cc96d2554230f90338b163bc6380a2a85", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 32, "original_max_position_embeddings": 4096, "pad_token_id": 32000, "resid_pdrop": 0.0, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000.0, "sliding_window": 2047, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 32064}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/phi3/microsoft/Phi-3-mini-4k-instruct/f6d3183f58f8544aad0f.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["Phi3ForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "auto_map": {"AutoConfig": "microsoft/Phi-3-mini-4k-instruct--configuration_phi3.Phi3Config", "AutoModelForCausalLM": "microsoft/Phi-3-mini-4k-instruct--modeling_phi3.Phi3ForCausalLM"}, "bos_token_id": 1, "embd_pdrop": 0.0, "eos_token_id": 32000, "hidden_act": "silu", "hidden_size": 3072, "initializer_range": 0.02, "intermediate_size": 8192, "max_position_embeddings": 4096, "model_type": "phi3", "neuron": {"auto_cast_type": "bf16", "batch_size": 1, "checkpoint_id": "microsoft/Phi-3-mini-4k-instruct", "checkpoint_revision": "0a67737cc96d2554230f90338b163bc6380a2a85", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 4, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 32, "original_max_position_embeddings": 4096, "pad_token_id": 32000, "resid_pdrop": 0.0, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000.0, "sliding_window": 2047, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 32064}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/phi3/yujiepan/phi-4-tiny-random/33891c5b973ec2bfb226.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["Phi3ForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "auto_map": {}, "bos_token_id": 100257, "embd_pdrop": 0.0, "eos_token_id": 100257, "hidden_act": "silu", "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 32, "max_position_embeddings": 16384, "model_type": "phi3", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "yujiepan/phi-4-tiny-random", "checkpoint_revision": "18a9a1168dc97ac6d128f811925670c275610f5a", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 2, "num_hidden_layers": 2, "num_key_value_heads": 1, "original_max_position_embeddings": 16384, "pad_token_id": 100257, "resid_pdrop": 0.0, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 250000, "sliding_window": null, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 100352}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/phi3/yujiepan/phi-4-tiny-random/48d7f1c1e412ea84b60e.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["Phi3ForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "auto_map": {}, "bos_token_id": 100257, "embd_pdrop": 0.0, "eos_token_id": 100257, "hidden_act": "silu", "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 32, "max_position_embeddings": 16384, "model_type": "phi3", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "yujiepan/phi-4-tiny-random", "checkpoint_revision": "18a9a1168dc97ac6d128f811925670c275610f5a", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 2, "num_hidden_layers": 2, "num_key_value_heads": 1, "original_max_position_embeddings": 16384, "pad_token_id": 100257, "resid_pdrop": 0.0, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 250000, "sliding_window": null, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 100352}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/phi3/yujiepan/phi-4-tiny-random/a37d77e276a4a0f5c4e8.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["Phi3ForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "auto_map": {}, "bos_token_id": 100257, "embd_pdrop": 0.0, "eos_token_id": 100257, "hidden_act": "silu", "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 32, "max_position_embeddings": 16384, "model_type": "phi3", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "yujiepan/phi-4-tiny-random", "checkpoint_revision": "18a9a1168dc97ac6d128f811925670c275610f5a", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 2, "num_hidden_layers": 2, "num_key_value_heads": 1, "original_max_position_embeddings": 16384, "pad_token_id": 100257, "resid_pdrop": 0.0, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 250000, "sliding_window": null, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 100352}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/qwen2/Qwen/Qwen2.5-0.5B/08b82ea5dab18178ad7f.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["Qwen2ForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 151643, "eos_token_id": 151643, "hidden_act": "silu", "hidden_size": 896, "initializer_range": 0.02, "intermediate_size": 4864, "max_position_embeddings": 32768, "max_window_layers": 24, "model_type": "qwen2", "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "Qwen/Qwen2.5-0.5B", "checkpoint_revision": "060db6499f32faf8b98477b0a26969ef7d8b9987", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 14, "num_hidden_layers": 24, "num_key_value_heads": 2, "rms_norm_eps": 1e-06, "rope_scaling": null, "rope_theta": 1000000.0, "sliding_window": null, "tie_word_embeddings": true, "torch_dtype": "bfloat16", "use_cache": true, "use_mrope": false, "use_sliding_window": false, "vocab_size": 151936}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev3/inference/gpt2/gpt2/780a42b2f8414c10d661.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "architectures": ["GPT2LMHeadModel"], "attn_pdrop": 0.1, "bos_token_id": 50256, "embd_pdrop": 0.1, "eos_token_id": 50256, "initializer_range": 0.02, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 1024, "n_embd": 768, "n_head": 12, "n_inner": null, "n_layer": 12, "n_positions": 1024, "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "gpt2", "checkpoint_revision": "607a30d783dfa663caf39e06633721c8d4cfcd7e", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 1024, "task": "text-generation"}, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "task_specific_params": {"text-generation": {"do_sample": true, "max_length": 50}}, "use_cache": true, "vocab_size": 50257}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev3/inference/granite/ibm-granite/granite-3.1-2b-instruct/bf71032506a251068193.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["GraniteForCausalLM"], "attention_bias": false, "attention_dropout": 0.1, "attention_multiplier": 0.015625, "bos_token_id": 0, "embedding_multiplier": 12.0, "eos_token_id": 0, "hidden_act": "silu", "hidden_size": 2048, "initializer_range": 0.02, "intermediate_size": 8192, "logits_scaling": 8.0, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "granite", "neuron": {"auto_cast_type": "bf16", "batch_size": 4, "checkpoint_id": "ibm-granite/granite-3.1-2b-instruct", "checkpoint_revision": "374ef54e020a3ce208c65e96d6213922a87d8952", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 40, "num_key_value_heads": 8, "pad_token_id": 0, "residual_multiplier": 0.22, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 5000000.0, "tie_word_embeddings": true, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 49155}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev3/inference/llama/meta-llama/Llama-3.2-1B-Instruct/c19d11f088aafb7b1382.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": [128001, 128008, 128009], "head_dim": 64, "hidden_act": "silu", "hidden_size": 2048, "initializer_range": 0.02, "intermediate_size": 8192, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "bf16", "batch_size": 4, "checkpoint_id": "meta-llama/Llama-3.2-1B-Instruct", "checkpoint_revision": "9213176726f574b556790deb65791e0c5aa438b6", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 16, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 32.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": true, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 128256}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev3/inference/llama/unsloth/Llama-3.2-1B-Instruct/49d61e18927b113552d1.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": 128009, "head_dim": 64, "hidden_act": "silu", "hidden_size": 2048, "initializer_range": 0.02, "intermediate_size": 8192, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "unsloth/Llama-3.2-1B-Instruct", "checkpoint_revision": "9b58d4a36161a1e49ecf0a69d20b2736fef8e438", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 16, "num_key_value_heads": 8, "pad_token_id": 128004, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 32.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": true, "torch_dtype": "bfloat16", "unsloth_fixed": true, "use_cache": true, "vocab_size": 128256}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev3/inference/llama/unsloth/Llama-3.2-1B-Instruct/9d4dd76cfd7e72235824.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": 128009, "head_dim": 64, "hidden_act": "silu", "hidden_size": 2048, "initializer_range": 0.02, "intermediate_size": 8192, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "bf16", "batch_size": 2, "checkpoint_id": "unsloth/Llama-3.2-1B-Instruct", "checkpoint_revision": "9b58d4a36161a1e49ecf0a69d20b2736fef8e438", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 16, "num_key_value_heads": 8, "pad_token_id": 128004, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 32.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": true, "torch_dtype": "bfloat16", "unsloth_fixed": true, "use_cache": true, "vocab_size": 128256}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev3/inference/llama/unsloth/Llama-3.2-3B-Instruct/b6352a1b005c1b4fdbee.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": 128009, "head_dim": 128, "hidden_act": "silu", "hidden_size": 3072, "initializer_range": 0.02, "intermediate_size": 8192, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "bf16", "batch_size": 2, "checkpoint_id": "unsloth/Llama-3.2-3B-Instruct", "checkpoint_revision": "889fdd323e8975eec1e1e0fc821a2c818e6ad494", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 24, "num_hidden_layers": 28, "num_key_value_heads": 8, "pad_token_id": 128004, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 32.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": true, "torch_dtype": "bfloat16", "unsloth_fixed": true, "use_cache": true, "vocab_size": 128256}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev3/inference/mistral/optimum/mistral-1.1b-testing/6668357e5d6bb2f40fe7.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MistralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "head_dim": 64, "hidden_act": "silu", "hidden_size": 2048, "initializer_range": 0.02, "intermediate_size": 5632, "max_position_embeddings": 32768, "model_type": "mistral", "neuron": {"auto_cast_type": "bf16", "batch_size": 4, "checkpoint_id": "optimum/mistral-1.1b-testing", "checkpoint_revision": "ce03bc8d47dbd2c173ff65f3a8de1325ba724195", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 22, "num_key_value_heads": 4, "rms_norm_eps": 1e-05, "rope_theta": 1000000.0, "sliding_window": null, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev3/inference/mixtral/dacorvo/Mixtral-tiny/f574775a377747e5836b.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["MixtralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "head_dim": 32, "hidden_act": "silu", "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 3584, "max_position_embeddings": 1024, "model_type": "mixtral", "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "dacorvo/Mixtral-tiny", "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 1024, "task": "text-generation"}, "num_attention_heads": 32, "num_experts_per_tok": 2, "num_hidden_layers": 2, "num_key_value_heads": 8, "num_local_experts": 8, "output_router_logits": false, "rms_norm_eps": 1e-05, "rope_theta": 10000.0, "router_aux_loss_coef": 0.001, "router_jitter_noise": 0.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
|