Synchronizing local compiler cache.
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +58 -0
- neuronxcc-2.15.143.0+e39249ad/MODULE_1f6cff7b8c7ee3876236+39f12043/model.neff +1 -1
- neuronxcc-2.15.143.0+e39249ad/MODULE_2eeecff6f903dd1aaf30+39f12043/model.neff +1 -1
- neuronxcc-2.15.143.0+e39249ad/MODULE_2fafc1ff132d5f941aad+39f12043/model.neff +1 -1
- neuronxcc-2.15.143.0+e39249ad/MODULE_3af7e7ef3791e98d3c6d+39f12043/model.neff +1 -1
- neuronxcc-2.15.143.0+e39249ad/MODULE_46e098601535b4170185+39f12043/model.neff +1 -1
- neuronxcc-2.15.143.0+e39249ad/MODULE_4f8465ac364975d34bcb+39f12043/model.neff +1 -1
- neuronxcc-2.15.143.0+e39249ad/MODULE_640ad25f4ab9e0ce4ac7+39f12043/model.neff +1 -1
- neuronxcc-2.15.143.0+e39249ad/MODULE_722d3c4ebc79e2bbd531+39f12043/model.neff +1 -1
- neuronxcc-2.15.143.0+e39249ad/MODULE_b420f7ff165f4f0517c3+39f12043/model.neff +1 -1
- neuronxcc-2.15.143.0+e39249ad/MODULE_bb86f1f35879e84a7798+39f12043/model.neff +1 -1
- neuronxcc-2.15.143.0+e39249ad/MODULE_f46c84f4be3f40eafec4+39f12043/model.neff +1 -1
- neuronxcc-2.15.143.0+e39249ad/MODULE_f63ff654073837167064+39f12043/model.neff +1 -1
- neuronxcc-2.15.143.0+e39249ad/MODULE_fcdf5b6858bafa453910+39f12043/model.neff +1 -1
- neuronxcc-2.15.143.0+e39249ad/MODULE_ff275b5606db64ceaea0+39f12043/model.neff +2 -2
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/granite/ibm-granite/granite-3.1-2b-instruct/0211503669c9e68fbbde.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/llama/meta-llama/Llama-3.1-8B-Instruct/25395edbfe4cb40ed4e9.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/llama/meta-llama/Llama-3.2-1B-Instruct/7a87208395f3db5ba7be.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/llama/meta-llama/Llama-3.2-1B-Instruct/8cd73c74d69aa6440798.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/llama/meta-llama/Llama-3.2-1B-Instruct/ebbf84c3779811da8d2c.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/llama/meta-llama/Meta-Llama-3.1-8B-Instruct/3b6b0065e24ca9d2eeb0.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/llama/princeton-nlp/Sheared-LLaMA-1.3B/8f3a6f200be91cc32a57.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/microsoft/Phi-3-medium-4k-instruct/ae83aa0fea9a9d0315d5.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/microsoft/Phi-3-mini-4k-instruct/13bb5f11971eace7eeed.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/microsoft/Phi-3-mini-4k-instruct/29904a0c0edb4679ee76.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/microsoft/Phi-3-mini-4k-instruct/9a957125ff99f484a66e.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/microsoft/Phi-3-mini-4k-instruct/d7892042a1da5eaa0020.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/microsoft/Phi-4-mini-instruct/6ff6f9ae27117fd24bd7.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/microsoft/phi-4/42644c76f78cfd9108b4.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/microsoft/phi-4/452310f61d209c411ec5.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/microsoft/phi-4/6faff6d0347b9b3197ed.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/microsoft/phi-4/aefe5e89e89e5ba40a59.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/microsoft/phi-4/b05887d807c6f53298d2.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/microsoft/phi-4/c488c1faf1a22992a2ea.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/microsoft/phi-4/e21879dbd09a21fb7a2e.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/microsoft/phi-4/f6232f240b5b98678a51.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/yujiepan/phi-4-tiny-random/33891c5b973ec2bfb226.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/yujiepan/phi-4-tiny-random/48d7f1c1e412ea84b60e.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/yujiepan/phi-4-tiny-random/a37d77e276a4a0f5c4e8.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/qwen2/Qwen/Qwen2.5-Math-1.5B/a5b11f2dd182bd813171.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/058e3ead348125e7808b.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/6bf9ef24cedd2d181630.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/7a4c62323b876a6ad082.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/gpt2/gpt2/780a42b2f8414c10d661.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/gpt2/hf-internal-testing/tiny-random-gpt2/3033d959070fc853146f.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/gpt2/hf-internal-testing/tiny-random-gpt2/46dc72bec418e48dc0b2.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/gpt2/hf-internal-testing/tiny-random-gpt2/fdfe8a25cf9f41629cb9.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/llama/llamafactory/tiny-random-Llama-3/2e8cb42c92e334dfbaa9.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/llama/llamafactory/tiny-random-Llama-3/5433fffb45752137c8de.json +1 -0
- neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/llama/llamafactory/tiny-random-Llama-3/8c48d1da1325809763e8.json +1 -0
.gitattributes
CHANGED
@@ -1867,3 +1867,61 @@ neuronxcc-2.15.143.0+e39249ad/MODULE_63cbd319a3d911724981+39f12043/model.neff fi
|
|
1867 |
neuronxcc-2.15.143.0+e39249ad/MODULE_8938e8bef94bcc33f561+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
1868 |
neuronxcc-2.15.143.0+e39249ad/MODULE_733a6bca18aa9a6325aa+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
1869 |
neuronxcc-2.15.143.0+e39249ad/MODULE_967d83f73b8fe30e4d14+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1867 |
neuronxcc-2.15.143.0+e39249ad/MODULE_8938e8bef94bcc33f561+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
1868 |
neuronxcc-2.15.143.0+e39249ad/MODULE_733a6bca18aa9a6325aa+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
1869 |
neuronxcc-2.15.143.0+e39249ad/MODULE_967d83f73b8fe30e4d14+39f12043/model.neff filter=lfs diff=lfs merge=lfs -text
|
1870 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_008512cdad534106824b+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1871 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_0d5e2c0e852741a3814c+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1872 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_10bd954bb3f712b9cf93+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1873 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_17de9f35d8b452d30d64+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1874 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_194792ca42c74e6c357a+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1875 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_215d5f9f03ff7f796870+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1876 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_2e2fefac0330c92cf935+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1877 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_38847bfec695e36ba44b+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1878 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_399d134629eb34a00d43+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1879 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_3aa61c04bd8d9ca8722a+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1880 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_40c932379328aee8b619+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1881 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_41bd479e657e66f398e1+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1882 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_45ee942ce275eb459e1d+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1883 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_4723142e6e2041791232+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1884 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_478acb1b0aae0b56f401+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1885 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_48bfe9ceb9631fdca2d4+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1886 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_48c6444adcac1b53fda6+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1887 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_4ed4eafdd8b57b2a58ba+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1888 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_51c384f52e045bea30bd+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1889 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_52108672c0a9ba2de711+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1890 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_5b45c246024901dcd434+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1891 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_626306e0d30994029849+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1892 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_66e776ad8f8401a7fa1f+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1893 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_6c9037784aa88d983971+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1894 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_7323a70130bb87727b84+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1895 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_7418d879b0b0dbe9d053+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1896 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_767b99765e404104c7bb+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1897 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_7f54d4d3885e2a479ab4+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1898 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_827f458fd58fe6442166+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1899 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_838aaca91716f9f756f4+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1900 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_88098bf708a0bb5dbef8/text_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
|
1901 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_88098bf708a0bb5dbef8/unet/model.neuron filter=lfs diff=lfs merge=lfs -text
|
1902 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_88098bf708a0bb5dbef8/vae_decoder/model.neuron filter=lfs diff=lfs merge=lfs -text
|
1903 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_88098bf708a0bb5dbef8/vae_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
|
1904 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_8d1befe3d92435e31fdf+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1905 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_922ce3591b342750be6b+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1906 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_949152ff315aeb616749+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1907 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_94932be571ca32c0d6c9+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1908 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_94df5568a6ee6b8f721b+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1909 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_95f239593a81a9baed26+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1910 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_9dc8af288609e8374339+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1911 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_9e3cdf817e494ba86fd4+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1912 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_a307a86e0d389d7d8d9e+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1913 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_a9bcc20735f75ac1f46e+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1914 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_b2821092d30cdeb76d6a+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1915 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_b2aaa317ae645625c9d1+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1916 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_b3ac85ee34a5d9e564ba+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1917 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_b4a7a4856dcbd900d6ed+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1918 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_c0740fe961da69526f92+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1919 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_cb16b651ea9d180d5cfd+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1920 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_d07d2aeb772cbb9c0ccd+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1921 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_de8368a717cfd6dfec57+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1922 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_e0765cf6df2204e3664e+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1923 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_e40769e5a35d3e5dacff+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1924 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_eca56defe775a7981c4b+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1925 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_ef26a928b1a6003d9b67+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1926 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_f65495564d03f1c90e3a+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
1927 |
+
neuronxcc-2.16.372.0+4a9b2326/MODULE_f914d922823d8ceace3c+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
|
neuronxcc-2.15.143.0+e39249ad/MODULE_1f6cff7b8c7ee3876236+39f12043/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3124224
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:51bf171f94f9804c13f8b2c5e2dfa8418ba7295354b59ec1d137e594fa7c4de2
|
3 |
size 3124224
|
neuronxcc-2.15.143.0+e39249ad/MODULE_2eeecff6f903dd1aaf30+39f12043/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 5264384
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3142a5e8019c33a5ef1b9cc53abb6dff4f8f1d50d0e04d749b04c81e8e78f6ef
|
3 |
size 5264384
|
neuronxcc-2.15.143.0+e39249ad/MODULE_2fafc1ff132d5f941aad+39f12043/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 10845184
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8ecf3dd4fd7bfcd84755db1c6e02e0c8636166a781de06fe106d75e2bf45267d
|
3 |
size 10845184
|
neuronxcc-2.15.143.0+e39249ad/MODULE_3af7e7ef3791e98d3c6d+39f12043/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4199424
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:335b430e4841d72c9841f63f75d7d0de45ca610c707c09b557a29554761e19c9
|
3 |
size 4199424
|
neuronxcc-2.15.143.0+e39249ad/MODULE_46e098601535b4170185+39f12043/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3523584
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d05301e480562da305f10a43be768d89386d1b48a8c84340c28f232b45e86fa6
|
3 |
size 3523584
|
neuronxcc-2.15.143.0+e39249ad/MODULE_4f8465ac364975d34bcb+39f12043/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3113984
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:90c4f57defdfb2564d5a6e55524dd2bc6096ebb019a82660921be2fbb57bcc5c
|
3 |
size 3113984
|
neuronxcc-2.15.143.0+e39249ad/MODULE_640ad25f4ab9e0ce4ac7+39f12043/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3339264
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:aed8feb1ed1fe46eb30eadb7c354481d90449089639f2116acd11c72c30ad0b2
|
3 |
size 3339264
|
neuronxcc-2.15.143.0+e39249ad/MODULE_722d3c4ebc79e2bbd531+39f12043/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3564544
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3fbaf06d4ad87a3f53dc034df20eb74f7a3468d54689007ca059c38b29f145a1
|
3 |
size 3564544
|
neuronxcc-2.15.143.0+e39249ad/MODULE_b420f7ff165f4f0517c3+39f12043/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3615744
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3a873b6981fc26e2572db4d3e64bb2b31d4ef36aeccf87b8c6c8620e6ef8351f
|
3 |
size 3615744
|
neuronxcc-2.15.143.0+e39249ad/MODULE_bb86f1f35879e84a7798+39f12043/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 24187904
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f90139cd0d7490783c7308c8f812756f7a897844edfa675eda4a16bdb6968985
|
3 |
size 24187904
|
neuronxcc-2.15.143.0+e39249ad/MODULE_f46c84f4be3f40eafec4+39f12043/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3800064
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a44f7412f20757c4f65770c17f04be082d44549d984ce2d9da57001e84042d6c
|
3 |
size 3800064
|
neuronxcc-2.15.143.0+e39249ad/MODULE_f63ff654073837167064+39f12043/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3206144
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:222d03619526335d9f9c03db8bc7234d21a6c33bbfa89ff0ea9d2d66866b0eb7
|
3 |
size 3206144
|
neuronxcc-2.15.143.0+e39249ad/MODULE_fcdf5b6858bafa453910+39f12043/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4680704
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cdb285f617fca099f1409e397c39eb94b3d01afc3677fdf3a460f37ad834bc4b
|
3 |
size 4680704
|
neuronxcc-2.15.143.0+e39249ad/MODULE_ff275b5606db64ceaea0+39f12043/model.neff
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:46d1f1d8dc3c1709dcb3ae0c6aac64f95a3524aa66a474765c06c6fad367de2c
|
3 |
+
size 3073024
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/granite/ibm-granite/granite-3.1-2b-instruct/0211503669c9e68fbbde.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["GraniteForCausalLM"], "attention_bias": false, "attention_dropout": 0.1, "attention_multiplier": 0.015625, "bos_token_id": 0, "embedding_multiplier": 12.0, "eos_token_id": 0, "hidden_act": "silu", "hidden_size": 2048, "initializer_range": 0.02, "intermediate_size": 8192, "logits_scaling": 8.0, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "granite", "neuron": {"auto_cast_type": "bf16", "batch_size": 2, "checkpoint_id": "ibm-granite/granite-3.1-2b-instruct", "checkpoint_revision": "374ef54e020a3ce208c65e96d6213922a87d8952", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 40, "num_key_value_heads": 8, "pad_token_id": 0, "residual_multiplier": 0.22, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 5000000.0, "tie_word_embeddings": true, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 49155}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/llama/meta-llama/Llama-3.1-8B-Instruct/25395edbfe4cb40ed4e9.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": [128001, 128008, 128009], "head_dim": 128, "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 14336, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "bf16", "batch_size": 4, "checkpoint_id": "meta-llama/Llama-3.1-8B-Instruct", "checkpoint_revision": "0e9e39f249a16976918f6564b8830bc894c89659", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 8.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 128256}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/llama/meta-llama/Llama-3.2-1B-Instruct/7a87208395f3db5ba7be.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": [128001, 128008, 128009], "head_dim": 64, "hidden_act": "silu", "hidden_size": 2048, "initializer_range": 0.02, "intermediate_size": 8192, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "bf16", "batch_size": 4, "checkpoint_id": "meta-llama/Llama-3.2-1B-Instruct", "checkpoint_revision": "9213176726f574b556790deb65791e0c5aa438b6", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 16, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 32.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": true, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 128256}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/llama/meta-llama/Llama-3.2-1B-Instruct/8cd73c74d69aa6440798.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": [128001, 128008, 128009], "head_dim": 64, "hidden_act": "silu", "hidden_size": 2048, "initializer_range": 0.02, "intermediate_size": 8192, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "bf16", "batch_size": 2, "checkpoint_id": "meta-llama/Llama-3.2-1B-Instruct", "checkpoint_revision": "9213176726f574b556790deb65791e0c5aa438b6", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 16, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 32.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": true, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 128256}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/llama/meta-llama/Llama-3.2-1B-Instruct/ebbf84c3779811da8d2c.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": [128001, 128008, 128009], "head_dim": 64, "hidden_act": "silu", "hidden_size": 2048, "initializer_range": 0.02, "intermediate_size": 8192, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "bf16", "batch_size": 1, "checkpoint_id": "meta-llama/Llama-3.2-1B-Instruct", "checkpoint_revision": "9213176726f574b556790deb65791e0c5aa438b6", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 16, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 32.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": true, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 128256}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/llama/meta-llama/Meta-Llama-3.1-8B-Instruct/3b6b0065e24ca9d2eeb0.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": [128001, 128008, 128009], "head_dim": 128, "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 14336, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "bf16", "batch_size": 1, "checkpoint_id": "meta-llama/Meta-Llama-3.1-8B-Instruct", "checkpoint_revision": "0e9e39f249a16976918f6564b8830bc894c89659", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 8.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 128256}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/llama/princeton-nlp/Sheared-LLaMA-1.3B/8f3a6f200be91cc32a57.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "head_dim": 128, "hidden_act": "silu", "hidden_size": 2048, "initializer_range": 0.02, "intermediate_size": 5504, "max_position_embeddings": 4096, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "princeton-nlp/Sheared-LLaMA-1.3B", "checkpoint_revision": "a4b76938edbf571ea7d7d9904861cbdca08809b4", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 16, "num_hidden_layers": 24, "num_key_value_heads": 16, "pad_token_id": 0, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float32", "use_cache": true, "vocab_size": 32000}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/microsoft/Phi-3-medium-4k-instruct/ae83aa0fea9a9d0315d5.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["Phi3ForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "auto_map": {"AutoConfig": "microsoft/Phi-3-medium-4k-instruct--configuration_phi3.Phi3Config", "AutoModelForCausalLM": "microsoft/Phi-3-medium-4k-instruct--modeling_phi3.Phi3ForCausalLM"}, "bos_token_id": 1, "embd_pdrop": 0.0, "eos_token_id": 32000, "hidden_act": "silu", "hidden_size": 5120, "initializer_range": 0.02, "intermediate_size": 17920, "max_position_embeddings": 4096, "model_type": "phi3", "neuron": {"auto_cast_type": "bf16", "batch_size": 4, "checkpoint_id": "microsoft/Phi-3-medium-4k-instruct", "checkpoint_revision": "ae004ae82eb6eddc32906dfacb1d6dfea8f91996", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 40, "num_hidden_layers": 40, "num_key_value_heads": 10, "original_max_position_embeddings": 4096, "pad_token_id": 32000, "partial_rotary_factor": 1.0, "resid_pdrop": 0.0, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000.0, "sliding_window": 2047, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 32064}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/microsoft/Phi-3-mini-4k-instruct/13bb5f11971eace7eeed.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["Phi3ForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "auto_map": {"AutoConfig": "microsoft/Phi-3-mini-4k-instruct--configuration_phi3.Phi3Config", "AutoModelForCausalLM": "microsoft/Phi-3-mini-4k-instruct--modeling_phi3.Phi3ForCausalLM"}, "bos_token_id": 1, "embd_pdrop": 0.0, "eos_token_id": 32000, "hidden_act": "silu", "hidden_size": 3072, "initializer_range": 0.02, "intermediate_size": 8192, "max_position_embeddings": 4096, "model_type": "phi3", "neuron": {"auto_cast_type": "bf16", "batch_size": 1, "checkpoint_id": "microsoft/Phi-3-mini-4k-instruct", "checkpoint_revision": "0a67737cc96d2554230f90338b163bc6380a2a85", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 32, "original_max_position_embeddings": 4096, "pad_token_id": 32000, "resid_pdrop": 0.0, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000.0, "sliding_window": 2047, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 32064}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/microsoft/Phi-3-mini-4k-instruct/29904a0c0edb4679ee76.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["Phi3ForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "auto_map": {"AutoConfig": "microsoft/Phi-3-mini-4k-instruct--configuration_phi3.Phi3Config", "AutoModelForCausalLM": "microsoft/Phi-3-mini-4k-instruct--modeling_phi3.Phi3ForCausalLM"}, "bos_token_id": 1, "embd_pdrop": 0.0, "eos_token_id": 32000, "hidden_act": "silu", "hidden_size": 3072, "initializer_range": 0.02, "intermediate_size": 8192, "max_position_embeddings": 4096, "model_type": "phi3", "neuron": {"auto_cast_type": "bf16", "batch_size": 4, "checkpoint_id": "microsoft/Phi-3-mini-4k-instruct", "checkpoint_revision": "0a67737cc96d2554230f90338b163bc6380a2a85", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 32, "original_max_position_embeddings": 4096, "pad_token_id": 32000, "resid_pdrop": 0.0, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000.0, "sliding_window": 2047, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 32064}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/microsoft/Phi-3-mini-4k-instruct/9a957125ff99f484a66e.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["Phi3ForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "auto_map": {"AutoConfig": "microsoft/Phi-3-mini-4k-instruct--configuration_phi3.Phi3Config", "AutoModelForCausalLM": "microsoft/Phi-3-mini-4k-instruct--modeling_phi3.Phi3ForCausalLM"}, "bos_token_id": 1, "embd_pdrop": 0.0, "eos_token_id": 32000, "hidden_act": "silu", "hidden_size": 3072, "initializer_range": 0.02, "intermediate_size": 8192, "max_position_embeddings": 4096, "model_type": "phi3", "neuron": {"auto_cast_type": "bf16", "batch_size": 4, "checkpoint_id": "microsoft/Phi-3-mini-4k-instruct", "checkpoint_revision": "0a67737cc96d2554230f90338b163bc6380a2a85", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 32, "original_max_position_embeddings": 4096, "pad_token_id": 32000, "partial_rotary_factor": 1.0, "resid_pdrop": 0.0, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000.0, "sliding_window": 2047, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 32064}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/microsoft/Phi-3-mini-4k-instruct/d7892042a1da5eaa0020.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["Phi3ForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "auto_map": {"AutoConfig": "microsoft/Phi-3-mini-4k-instruct--configuration_phi3.Phi3Config", "AutoModelForCausalLM": "microsoft/Phi-3-mini-4k-instruct--modeling_phi3.Phi3ForCausalLM"}, "bos_token_id": 1, "embd_pdrop": 0.0, "eos_token_id": 32000, "hidden_act": "silu", "hidden_size": 3072, "initializer_range": 0.02, "intermediate_size": 8192, "max_position_embeddings": 4096, "model_type": "phi3", "neuron": {"auto_cast_type": "bf16", "batch_size": 2, "checkpoint_id": "microsoft/Phi-3-mini-4k-instruct", "checkpoint_revision": "0a67737cc96d2554230f90338b163bc6380a2a85", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 32, "original_max_position_embeddings": 4096, "pad_token_id": 32000, "resid_pdrop": 0.0, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000.0, "sliding_window": 2047, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 32064}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/microsoft/Phi-4-mini-instruct/6ff6f9ae27117fd24bd7.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["Phi3ForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "auto_map": {"AutoConfig": "microsoft/Phi-4-mini-instruct--configuration_phi3.Phi3Config", "AutoModelForCausalLM": "microsoft/Phi-4-mini-instruct--modeling_phi3.Phi3ForCausalLM", "AutoTokenizer": "microsoft/Phi-4-mini-instruct--Xenova/gpt-4o"}, "bos_token_id": 199999, "embd_pdrop": 0.0, "eos_token_id": 199999, "full_attn_mod": 1, "hidden_act": "silu", "hidden_size": 3072, "initializer_range": 0.02, "intermediate_size": 8192, "interpolate_factor": 1, "lm_head_bias": false, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "phi3", "neuron": {"auto_cast_type": "bf16", "batch_size": 4, "checkpoint_id": "microsoft/Phi-4-mini-instruct", "checkpoint_revision": "4b00ec8714b0cb224e4fb33380cbf0919f177f3e", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 24, "num_hidden_layers": 32, "num_key_value_heads": 8, "original_max_position_embeddings": 4096, "pad_token_id": 199999, "partial_rotary_factor": 0.75, "resid_pdrop": 0.0, "rms_norm_eps": 1e-05, "rope_scaling": {"long_factor": [1, 1.118320672, 1.250641126, 1.398617824, 1.564103225, 1.74916897, 1.956131817, 2.187582649, 2.446418898, 2.735880826, 3.059592084, 3.421605075, 3.826451687, 4.279200023, 4.785517845, 5.351743533, 5.984965424, 6.693110555, 7.485043894, 8.370679318, 9.36110372, 10.4687158, 11.70738129, 13.09260651, 14.64173252, 16.37415215, 18.31155283, 20.47818807, 22.90118105, 25.61086418, 28.64115884, 32.03, 32.1, 32.13, 32.23, 32.6, 32.61, 32.64, 32.66, 32.7, 32.71, 32.93, 32.97, 33.28, 33.49, 33.5, 44.16, 47.77], "short_factor": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "type": "longrope"}, "rope_theta": 10000.0, "sliding_window": 262144, "tie_word_embeddings": true, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 200064}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/microsoft/phi-4/42644c76f78cfd9108b4.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["Phi3ForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 100257, "embd_pdrop": 0.0, "eos_token_id": 100265, "hidden_act": "silu", "hidden_size": 5120, "initializer_range": 0.02, "intermediate_size": 17920, "max_position_embeddings": 16384, "model_type": "phi3", "neuron": {"auto_cast_type": "bf16", "batch_size": 1, "checkpoint_id": "microsoft/phi-4", "checkpoint_revision": "187ef0342fff0eb3333be9f00389385e95ef0b61", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 40, "num_hidden_layers": 40, "num_key_value_heads": 10, "original_max_position_embeddings": 16384, "pad_token_id": 100349, "partial_rotary_factor": 1.0, "resid_pdrop": 0.0, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 250000, "sliding_window": null, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 100352}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/microsoft/phi-4/452310f61d209c411ec5.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["Phi3ForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 100257, "embd_pdrop": 0.0, "eos_token_id": 100265, "hidden_act": "silu", "hidden_size": 5120, "initializer_range": 0.02, "intermediate_size": 17920, "max_position_embeddings": 16384, "model_type": "phi3", "neuron": {"auto_cast_type": "bf16", "batch_size": 1, "checkpoint_id": "microsoft/phi-4", "checkpoint_revision": "187ef0342fff0eb3333be9f00389385e95ef0b61", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 40, "num_hidden_layers": 40, "num_key_value_heads": 10, "original_max_position_embeddings": 16384, "pad_token_id": 100349, "resid_pdrop": 0.0, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 250000, "sliding_window": null, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 100352}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/microsoft/phi-4/6faff6d0347b9b3197ed.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["Phi3ForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 100257, "embd_pdrop": 0.0, "eos_token_id": 100265, "hidden_act": "silu", "hidden_size": 5120, "initializer_range": 0.02, "intermediate_size": 17920, "max_position_embeddings": 16384, "model_type": "phi3", "neuron": {"auto_cast_type": "bf16", "batch_size": 2, "checkpoint_id": "microsoft/phi-4", "checkpoint_revision": "187ef0342fff0eb3333be9f00389385e95ef0b61", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 40, "num_hidden_layers": 40, "num_key_value_heads": 10, "original_max_position_embeddings": 16384, "pad_token_id": 100349, "resid_pdrop": 0.0, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 250000, "sliding_window": null, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 100352}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/microsoft/phi-4/aefe5e89e89e5ba40a59.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["Phi3ForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 100257, "embd_pdrop": 0.0, "eos_token_id": 100265, "hidden_act": "silu", "hidden_size": 5120, "initializer_range": 0.02, "intermediate_size": 17920, "max_position_embeddings": 16384, "model_type": "phi3", "neuron": {"auto_cast_type": "bf16", "batch_size": 24, "checkpoint_id": "microsoft/phi-4", "checkpoint_revision": "187ef0342fff0eb3333be9f00389385e95ef0b61", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 10, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 40, "num_hidden_layers": 40, "num_key_value_heads": 10, "original_max_position_embeddings": 16384, "pad_token_id": 100349, "resid_pdrop": 0.0, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 250000, "sliding_window": null, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 100352}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/microsoft/phi-4/b05887d807c6f53298d2.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["Phi3ForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 100257, "embd_pdrop": 0.0, "eos_token_id": 100265, "hidden_act": "silu", "hidden_size": 5120, "initializer_range": 0.02, "intermediate_size": 17920, "max_position_embeddings": 16384, "model_type": "phi3", "neuron": {"auto_cast_type": "bf16", "batch_size": 16, "checkpoint_id": "microsoft/phi-4", "checkpoint_revision": "187ef0342fff0eb3333be9f00389385e95ef0b61", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 40, "num_hidden_layers": 40, "num_key_value_heads": 10, "original_max_position_embeddings": 16384, "pad_token_id": 100349, "resid_pdrop": 0.0, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 250000, "sliding_window": null, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 100352}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/microsoft/phi-4/c488c1faf1a22992a2ea.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["Phi3ForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 100257, "embd_pdrop": 0.0, "eos_token_id": 100265, "hidden_act": "silu", "hidden_size": 5120, "initializer_range": 0.02, "intermediate_size": 17920, "max_position_embeddings": 16384, "model_type": "phi3", "neuron": {"auto_cast_type": "bf16", "batch_size": 16, "checkpoint_id": "microsoft/phi-4", "checkpoint_revision": "187ef0342fff0eb3333be9f00389385e95ef0b61", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 40, "num_hidden_layers": 40, "num_key_value_heads": 10, "original_max_position_embeddings": 16384, "pad_token_id": 100349, "partial_rotary_factor": 1.0, "resid_pdrop": 0.0, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 250000, "sliding_window": null, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 100352}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/microsoft/phi-4/e21879dbd09a21fb7a2e.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["Phi3ForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 100257, "embd_pdrop": 0.0, "eos_token_id": 100265, "hidden_act": "silu", "hidden_size": 5120, "initializer_range": 0.02, "intermediate_size": 17920, "max_position_embeddings": 16384, "model_type": "phi3", "neuron": {"auto_cast_type": "bf16", "batch_size": 8, "checkpoint_id": "microsoft/phi-4", "checkpoint_revision": "187ef0342fff0eb3333be9f00389385e95ef0b61", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 10, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 40, "num_hidden_layers": 40, "num_key_value_heads": 10, "original_max_position_embeddings": 16384, "pad_token_id": 100349, "resid_pdrop": 0.0, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 250000, "sliding_window": null, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 100352}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/microsoft/phi-4/f6232f240b5b98678a51.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["Phi3ForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 100257, "embd_pdrop": 0.0, "eos_token_id": 100265, "hidden_act": "silu", "hidden_size": 5120, "initializer_range": 0.02, "intermediate_size": 17920, "max_position_embeddings": 16384, "model_type": "phi3", "neuron": {"auto_cast_type": "bf16", "batch_size": 16, "checkpoint_id": "microsoft/phi-4", "checkpoint_revision": "187ef0342fff0eb3333be9f00389385e95ef0b61", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 12, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 40, "num_hidden_layers": 40, "num_key_value_heads": 10, "original_max_position_embeddings": 16384, "pad_token_id": 100349, "resid_pdrop": 0.0, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 250000, "sliding_window": null, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 100352}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/yujiepan/phi-4-tiny-random/33891c5b973ec2bfb226.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["Phi3ForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "auto_map": {}, "bos_token_id": 100257, "embd_pdrop": 0.0, "eos_token_id": 100257, "hidden_act": "silu", "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 32, "max_position_embeddings": 16384, "model_type": "phi3", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "yujiepan/phi-4-tiny-random", "checkpoint_revision": "18a9a1168dc97ac6d128f811925670c275610f5a", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 2, "num_hidden_layers": 2, "num_key_value_heads": 1, "original_max_position_embeddings": 16384, "pad_token_id": 100257, "resid_pdrop": 0.0, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 250000, "sliding_window": null, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 100352}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/yujiepan/phi-4-tiny-random/48d7f1c1e412ea84b60e.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["Phi3ForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "auto_map": {}, "bos_token_id": 100257, "embd_pdrop": 0.0, "eos_token_id": 100257, "hidden_act": "silu", "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 32, "max_position_embeddings": 16384, "model_type": "phi3", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "yujiepan/phi-4-tiny-random", "checkpoint_revision": "18a9a1168dc97ac6d128f811925670c275610f5a", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 2, "num_hidden_layers": 2, "num_key_value_heads": 1, "original_max_position_embeddings": 16384, "pad_token_id": 100257, "resid_pdrop": 0.0, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 250000, "sliding_window": null, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 100352}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/phi3/yujiepan/phi-4-tiny-random/a37d77e276a4a0f5c4e8.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["Phi3ForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "auto_map": {}, "bos_token_id": 100257, "embd_pdrop": 0.0, "eos_token_id": 100257, "hidden_act": "silu", "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 32, "max_position_embeddings": 16384, "model_type": "phi3", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "yujiepan/phi-4-tiny-random", "checkpoint_revision": "18a9a1168dc97ac6d128f811925670c275610f5a", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 2, "num_hidden_layers": 2, "num_key_value_heads": 1, "original_max_position_embeddings": 16384, "pad_token_id": 100257, "resid_pdrop": 0.0, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 250000, "sliding_window": null, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 100352}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev1/inference/qwen2/Qwen/Qwen2.5-Math-1.5B/a5b11f2dd182bd813171.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["Qwen2ForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 151643, "eos_token_id": 151643, "hidden_act": "silu", "hidden_size": 1536, "initializer_range": 0.02, "intermediate_size": 8960, "max_position_embeddings": 4096, "max_window_layers": 21, "model_type": "qwen2", "neuron": {"auto_cast_type": "bf16", "batch_size": 2, "checkpoint_id": "Qwen/Qwen2.5-Math-1.5B", "checkpoint_revision": "4a83ca6e4526a4f2da3aa259ec36c259f66b2ab2", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 12, "num_hidden_layers": 28, "num_key_value_heads": 2, "rms_norm_eps": 1e-06, "rope_scaling": null, "rope_theta": 10000, "sliding_window": null, "tie_word_embeddings": true, "torch_dtype": "bfloat16", "use_cache": true, "use_mrope": false, "use_sliding_window": false, "vocab_size": 151936}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/058e3ead348125e7808b.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/6bf9ef24cedd2d181630.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/7a4c62323b876a6ad082.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/gpt2/gpt2/780a42b2f8414c10d661.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "architectures": ["GPT2LMHeadModel"], "attn_pdrop": 0.1, "bos_token_id": 50256, "embd_pdrop": 0.1, "eos_token_id": 50256, "initializer_range": 0.02, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 1024, "n_embd": 768, "n_head": 12, "n_inner": null, "n_layer": 12, "n_positions": 1024, "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "gpt2", "checkpoint_revision": "607a30d783dfa663caf39e06633721c8d4cfcd7e", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 1024, "task": "text-generation"}, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "task_specific_params": {"text-generation": {"do_sample": true, "max_length": 50}}, "use_cache": true, "vocab_size": 50257}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/gpt2/hf-internal-testing/tiny-random-gpt2/3033d959070fc853146f.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/gpt2/hf-internal-testing/tiny-random-gpt2/46dc72bec418e48dc0b2.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/gpt2/hf-internal-testing/tiny-random-gpt2/fdfe8a25cf9f41629cb9.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/llama/llamafactory/tiny-random-Llama-3/2e8cb42c92e334dfbaa9.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": [128001, 128008, 128009], "head_dim": 4, "hidden_act": "silu", "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 64, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "llamafactory/tiny-random-Llama-3", "checkpoint_revision": "bf2a2e3bf199ad2ee96f02a3c00246c608db22a8", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 4, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 8.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 128256}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/llama/llamafactory/tiny-random-Llama-3/5433fffb45752137c8de.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": [128001, 128008, 128009], "head_dim": 4, "hidden_act": "silu", "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 64, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "llamafactory/tiny-random-Llama-3", "checkpoint_revision": "bf2a2e3bf199ad2ee96f02a3c00246c608db22a8", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 4, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 8.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 128256}
|
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/llama/llamafactory/tiny-random-Llama-3/8c48d1da1325809763e8.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": [128001, 128008, 128009], "head_dim": 4, "hidden_act": "silu", "hidden_size": 16, "initializer_range": 0.02, "intermediate_size": 64, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "llamafactory/tiny-random-Llama-3", "checkpoint_revision": "bf2a2e3bf199ad2ee96f02a3c00246c608db22a8", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 4, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 8.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 128256}
|