diff --git "a/perf-df-awq-1xA10.csv" "b/perf-df-awq-1xA10.csv" --- "a/perf-df-awq-1xA10.csv" +++ "b/perf-df-awq-1xA10.csv" @@ -67,7 +67,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e52b4-0bad542438dadd831ce8714a;ef2f8906-18f9-4c44-9f23-3c19c426f8da) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6671000d-428d47014afb2f0b6612d61a;cc2668c2-68d1-40be-924d-6d6f9b614c4d) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -166,7 +166,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e53ae-6a562b9a0722eecb2fa7743e;1e80040b-7d50-4cc2-8888-9c95f1117aa2) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710108-27d7c6755166edd06af205e6;9007585e-8249-49dd-b388-9d909328aea1) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -434,7 +434,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e43a8-38a384c65e6307bb493c15f6;5241a3fe-dcf8-45bd-aa6b-3010480cd34a) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e713-4ca24bbd242ee826122276e2;63a9df1c-df75-4d27-9aa1-daa34163176f) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -639,7 +639,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5306-1e7cbd9215be8acb6ffdd50c;ab860fa0-49ad-40b3-a08c-763b08f0d212) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710063-240cc2443aefa0cf42f45dbc;f208d84b-ca63-46df-953e-a6d6c4fa70e2) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -699,7 +699,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e4353-58f52b2f4dbebf641fd9cebb;c43192af-6448-47d7-a3b4-5881c02a93f7) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e6c0-11039e367626dbde0766e092;b31fe90e-bc86-4e94-a201-54bba9792ad0) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -1207,7 +1207,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e3e49-0759080462e429273e9bac0d;8f180fe4-fcaf-48eb-8e1e-29363e096123) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e17f-3898ceb8229c5a133838fa5b;428b803d-adc2-4f2e-99be-9c25139fe208) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -1292,7 +1292,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e51fc-488b5c3866978f401db7a39d;c49e96fd-2eb2-4d48-8b27-0ac787e9f831) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670ff3b-2e2479de1c88c44d4d740d29;cb2d5f39-728c-4e98-a749-c973ca1114b9) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1436,7 +1436,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e504d-52e5f8f601af569429fe03c9;e44d7a13-3ee6-4993-90cf-63c823371e15) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fd37-2fd265b201f6c6633fba3184;3955387c-efd5-4407-8174-25167c260578) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1608,7 +1608,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e54a8-1d304c911b77953845a1a839;f9552801-47ec-40a5-be61-b20890acc942) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710200-0b1da390011933f07093a2c8;bb4b3a22-3935-4fcd-806d-1f977b72eebd) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1707,7 +1707,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4ff8-385ab6421bdf60f579240014;b8004d37-032d-4467-9552-272a362aa46d) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fcd8-08a751852f1eb5683df2f28f;97288ec5-ccea-46d8-8248-f55077791c37) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1806,7 +1806,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5151-2e9da5936b09adc91c2ce2ff;baf891f2-4637-4416-9098-f5216355ec1c) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fe6e-47171ea03819c8904a6f8d59;4c02dd94-db8b-42fc-abca-b37a005f7c96) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1897,7 +1897,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e593a-6a544b6d5583e8cf66fd0dce;e381e8c9-92e6-4fb6-bed8-8f94493de30a) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667106af-115ce27a2382dcfa1c7abfda;285bfa31-b918-4e50-a3a5-2f3fee036ff2) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -1982,7 +1982,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e51a5-25f35b3c7aeef37845d16843;9e1d8b9d-cf33-4786-bf3a-0d7840049e06) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fed1-3f64ba1929ec998c2aa8bdb9;7e7e7ab3-e95b-4d8f-b418-0ace9d233a4a) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2072,7 +2072,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e4301-74611f480b42caed73ec868c;f481202f-39de-49e2-919b-0aea0e87aaa5) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e66c-78371b0117609d576856dbb0;1518c8a5-d0fb-406a-b4c2-62a493122d37) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -2255,7 +2255,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e50fd-1f50232609f9b52b74443a4f;44bcac88-f368-4628-8747-79cd273f8c83) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fdf7-344ee75f038872d766601347;9847bfbb-a12b-4b0e-ba77-14e25e208890) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2384,7 +2384,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e43fe-73437ea333f878fc24be9058;c4bd350e-893b-40be-8011-9925421be2b5) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e767-7127696a191030d90fcce079;ccdff8c0-6dd3-421f-a4e3-d8fcecd682dd) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2513,7 +2513,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5402-271cfb5f4991b5a64c8447c3;cf8ed820-c44f-490e-87c8-691bc94a2b7c) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6671015c-131ccc8c3b9adbfa6336d143;70e5314c-89e5-4807-a677-fd7d0b0729a5) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2613,7 +2613,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4d52-672ec95809e5574124c1cdde;0c0a7c2b-d3f9-4859-81f0-43b65ed53807) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670f9c4-1c21021b2e4724d554f83c8e;64d94c4d-63b9-448d-b6a6-8733e35a8b9c) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2922,39 +2922,7 @@ ValueError: OPTForCausalLM does not support an attention implementation through ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,microsoft/phi-1_5,microsoft/phi-1_5,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemm,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,MB,1485.733888,2103.967744,0.0,1457.52064,1272.750592,s,10,1.3342720336914062,0.13342720336914063,0.0017667378558565091,0.1333674545288086,0.13440001983642577,0.13625353012084962,0.13773633834838866,"[0.13810704040527344, 0.13359715270996095, 0.13159478759765625, 0.13398812866210938, 0.13163987731933594, 0.132114013671875, 0.13270646667480468, 0.1334988098144531, 0.13378965759277345, 0.13323609924316407]",tokens/s,1918.6492224658912,kWh,1.5465521064155546e-06,8.474274160367454e-07,6.439467633532405e-06,8.833447155984705e-06,tokens/kWh,28980758.641495775,MB,1485.733888,2103.967744,0.0,1457.52064,1369.423872,s,10,77.59171484375001,7.759171484375001,0.021550685509445954,7.750958740234375,7.78698193359375,7.79351220703125,7.798736425781249,"[7.7519140625, 7.7428955078125, 7.78553076171875, 7.781552734375, 7.80004248046875, 7.76520556640625, 7.75000341796875, 7.7375029296875, 7.73547216796875, 7.74159521484375]",tokens/s,8.119423591406116,kWh,9.154879443379708e-05,5.017541917467626e-05,0.00037131786234986455,0.0005130420759583379,tokens/kWh,122796.94581057319,,s,629,78.66502353668211,0.12506363042397795,0.0158250884711649,0.12294553375244141,0.1240623077392578,0.1244600341796875,0.2556754669189453,"[0.12330188751220703, 0.12316671752929688, 0.12365106964111328, 0.12313600158691407, 0.12294246673583985, 0.12248063659667968, 0.12371660614013671, 0.12260147094726563, 0.12286566162109375, 0.12282879638671874, 0.12315750122070312, 0.12320972442626953, 0.12280012512207031, 0.12268236541748047, 0.122640380859375, 0.12257997131347656, 0.12299161529541015, 0.1229834213256836, 0.12259737396240235, 0.12290866851806641, 0.12273561859130859, 0.12333164978027343, 0.12384864044189453, 0.12324454498291015, 0.12292301177978515, 0.12274278259277344, 0.1227489242553711, 0.12272640228271485, 0.12298239898681641, 0.12477235412597656, 0.12357427215576172, 0.12362035369873046, 0.12306233978271484, 0.12299974060058594, 0.12285030364990235, 0.12300287628173828, 0.12316060638427734, 0.1231595230102539, 0.12301516723632812, 0.12314726257324218, 0.12401152038574219, 0.12320358276367188, 0.12276121520996094, 0.12288307189941407, 0.12321382141113281, 0.12264653015136719, 0.12301721954345703, 0.12259737396240235, 0.12273664093017578, 0.122893310546875, 0.12342066955566407, 0.1232015380859375, 0.12298239898681641, 0.12279090881347657, 0.1230448989868164, 0.12328444671630859, 0.12287181091308594, 0.12288511657714844, 0.12308684539794922, 0.1227540512084961, 0.12283084869384765, 0.12267212677001953, 0.25649664306640624, 0.1227540512084961, 0.12280220794677735, 0.12263113403320312, 0.12264345550537109, 0.12267417907714843, 0.1228400650024414, 0.12283596801757812, 0.12282470703125, 0.12279603576660156, 0.12281446075439453, 0.12269875335693359, 0.12270694732666015, 0.12278681945800782, 0.12276121520996094, 0.12260044860839844, 0.12260659027099609, 0.12257791900634765, 0.12294041442871094, 0.1228779525756836, 0.12287999725341797, 0.12280831909179687, 0.12294451141357422, 0.12248985290527344, 0.12257484436035156, 0.1226055679321289, 0.12298137664794923, 0.12271308898925781, 0.12264447784423828, 0.12263935852050781, 0.12252365112304688, 0.12261990356445313, 0.12269977569580078, 0.12264345550537109, 0.12255129241943359, 0.12263833618164062, 0.12267212677001953, 0.12288102722167969, 0.12292915344238281, 0.12263935852050781, 0.12389888000488282, 0.12297216033935547, 0.12294553375244141, 0.12293427276611328, 0.12263526153564454, 0.12274585723876953, 0.12553011322021485, 0.12401561737060547, 0.12355072021484376, 0.12335718536376954, 0.12291584014892579, 0.12288819122314452, 0.12369510650634766, 0.12324147033691406, 0.12273667144775391, 0.12291478729248047, 0.12294758605957032, 0.12287385559082031, 0.12286669158935547, 0.12300697326660157, 0.12317391967773438, 0.12299874877929687, 0.1229271011352539, 0.25462066650390625, 0.12265164947509766, 0.1226270751953125, 0.12303667449951172, 0.12276227569580078, 0.12365821075439454, 0.12293939208984375, 0.12266496276855468, 0.12368895721435547, 0.12677529907226562, 0.12441395568847656, 0.12445696258544922, 0.12463513946533203, 0.12443033599853516, 0.12438224029541016, 0.12425520324707032, 0.12467916870117188, 0.12446310424804688, 0.12442931365966797, 0.12455731201171875, 0.12458188629150391, 0.12372480010986328, 0.12267417907714843, 0.12398899078369141, 0.12292813110351562, 0.12310527801513672, 0.12295891571044922, 0.12267308807373047, 0.12258099365234375, 0.12275609588623047, 0.12290764617919922, 0.12266086578369141, 0.12255948638916016, 0.12306329345703125, 0.12263116455078125, 0.12326604461669922, 0.12354764556884766, 0.12335718536376954, 0.12372691345214844, 0.12280210876464843, 0.12289638519287109, 0.12269055938720703, 0.12372787475585938, 0.124010498046875, 0.12365824127197265, 0.12331314849853516, 0.12313497924804688, 0.12537753295898438, 0.12376166534423828, 0.12374527740478515, 0.12372582244873047, 0.1231278076171875, 0.12385382080078125, 0.12404633331298828, 0.12378214263916015, 0.12360908508300782, 0.12295782470703125, 0.12494950103759765, 0.12373811340332032, 0.12298035430908204, 0.12348416137695313, 0.12417740631103516, 0.12380774688720703, 0.2588282775878906, 0.12419481658935547, 0.12356505584716797, 0.12310323333740235, 0.12305203247070312, 0.12337971496582031, 0.12367977905273438, 0.12359062194824219, 0.12260147094726563, 0.12353740692138672, 0.12354354858398438, 0.1232701416015625, 0.12446720123291016, 0.12432588958740234, 0.1239582748413086, 0.12362137603759765, 0.12360499572753907, 0.12407295989990234, 0.12389170837402344, 0.12347392272949219, 0.1235968017578125, 0.12413645172119141, 0.12480512237548828, 0.12312678527832031, 0.12333670043945312, 0.12412928009033203, 0.12416000366210937, 0.12297113800048828, 0.12327529907226563, 0.12358857727050782, 0.12392755126953126, 0.1234176025390625, 0.12362649536132812, 0.12375859069824219, 0.12381183624267578, 0.12392758178710937, 0.12367868804931641, 0.12318924713134766, 0.12382310485839844, 0.12388658905029297, 0.12348006439208985, 0.12391423797607422, 0.12419583892822265, 0.12329676818847657, 0.12479590606689453, 0.12373197174072266, 0.12389478302001954, 0.12393472290039062, 0.12317702484130859, 0.12292499542236328, 0.12298649597167968, 0.12279808044433593, 0.12290457916259766, 0.1228584976196289, 0.12289842987060547, 0.1226424331665039, 0.12311244964599609, 0.12252780914306641, 0.12262598419189454, 0.12282470703125, 0.12261888122558594, 0.12280217742919922, 0.12272025299072266, 0.25568666076660157, 0.1239900131225586, 0.12271616363525391, 0.12272128295898438, 0.1237401580810547, 0.12348108673095703, 0.12270182037353515, 0.12281241607666016, 0.12259839630126954, 0.12345139312744141, 0.12484710693359374, 0.123040771484375, 0.12316671752929688, 0.12355583953857421, 0.12351078033447266, 0.12384358215332031, 0.12355174255371094, 0.12602265930175782, 0.12446208190917969, 0.12428800201416015, 0.12392345428466797, 0.12333055877685548, 0.12432588958740234, 0.12359986877441406, 0.12368793487548828, 0.12390502166748046, 0.12386921691894531, 0.1236797103881836, 0.12345037078857422, 0.12340735626220703, 0.12414361572265625, 0.12356301116943359, 0.12358767700195313, 0.12459613037109375, 0.12439347076416016, 0.12301107025146485, 0.12367359924316407, 0.12390092468261718, 0.12395724487304688, 0.12516966247558595, 0.12354662322998047, 0.12346470642089843, 0.123863037109375, 0.12409139251708984, 0.12322509002685547, 0.12475084686279297, 0.12426445007324219, 0.12396953582763671, 0.12330086517333984, 0.12360908508300782, 0.12444057464599609, 0.12403609466552734, 0.12382208251953125, 0.1240606689453125, 0.12347494506835938, 0.12344012451171875, 0.12362751770019531, 0.1256468505859375, 0.12401663970947266, 0.12414873504638672, 0.12401868438720703, 0.12380364990234374, 0.12371561431884766, 0.25564668273925784, 0.12369203186035156, 0.12362137603759765, 0.12392345428466797, 0.12388044738769531, 0.12370022583007813, 0.12369715118408203, 0.12311347198486328, 0.12364492797851563, 0.123863037109375, 0.12393574523925781, 0.12324249267578125, 0.12262403106689453, 0.12281238555908203, 0.12295372772216796, 0.12290764617919922, 0.12269363403320313, 0.12262297821044922, 0.1231677474975586, 0.12281446075439453, 0.12263526153564454, 0.122714111328125, 0.12280524444580078, 0.12276838684082031, 0.1227110366821289, 0.12276735687255859, 0.12391014099121093, 0.12351795196533204, 0.12277555084228516, 0.12253388977050782, 0.12272025299072266, 0.12271622467041016, 0.12318611145019531, 0.12313190460205078, 0.12436070251464844, 0.12357734680175782, 0.12378726196289062, 0.12353024291992187, 0.12572364807128905, 0.12379545593261719, 0.12325888061523438, 0.12332441711425782, 0.12372991943359375, 0.12286566162109375, 0.1225134048461914, 0.12262400054931641, 0.12326297760009766, 0.12308582305908203, 0.12292403411865234, 0.1229148178100586, 0.12318924713134766, 0.12364492797851563, 0.12364291381835937, 0.12278781127929687, 0.12316159820556641, 0.12317593383789062, 0.12396851348876953, 0.12279193878173827, 0.12260761260986328, 0.12462694549560546, 0.12346470642089843, 0.12266598510742187, 0.1229148178100586, 0.2559385528564453, 0.12347289276123047, 0.12349951934814453, 0.12279500579833984, 0.12253695678710938, 0.1227325439453125, 0.12260454559326171, 0.12274380493164062, 0.12233113861083984, 0.12244992065429687, 0.12274790191650391, 0.12282572937011718, 0.12250214385986329, 0.12254208374023437, 0.12273458862304687, 0.12241919708251953, 0.12272844696044923, 0.1226250228881836, 0.12303974151611329, 0.1227663345336914, 0.12259839630126954, 0.12243353271484375, 0.12274380493164062, 0.12401561737060547, 0.12277452850341797, 0.1226844482421875, 0.12271612548828124, 0.12248268890380859, 0.12263935852050781, 0.12302540588378906, 0.12259327697753906, 0.1225697250366211, 0.12243456268310547, 0.12261273956298828, 0.12260352325439453, 0.12275507354736329, 0.12261068725585937, 0.12276428985595703, 0.12263942718505859, 0.12258604431152344, 0.12272230529785157, 0.12245708465576172, 0.12246527862548828, 0.12284928131103516, 0.12406886291503906, 0.12364492797851563, 0.12372889709472656, 0.12372480010986328, 0.1236131820678711, 0.12349132537841796, 0.12373094177246094, 0.12433715057373047, 0.12420198059082031, 0.12350566101074219, 0.12349030303955078, 0.12351795196533204, 0.1237022705078125, 0.12365004730224609, 0.12362035369873046, 0.12419891357421875, 0.12375347137451172, 0.12319641876220704, 0.12275917053222657, 0.2567250061035156, 0.1229854736328125, 0.12295680236816406, 0.1226270751953125, 0.12288614654541016, 0.12259225463867188, 0.12339711761474609, 0.1230387191772461, 0.12344627380371094, 0.12426547241210938, 0.12386918640136718, 0.12261785888671875, 0.12307660675048829, 0.1229639663696289, 0.12252467346191406, 0.12267314910888671, 0.12372172546386719, 0.12260659027099609, 0.12311961364746093, 0.1224427490234375, 0.12306841278076172, 0.12283699035644531, 0.12256665802001954, 0.12251955413818359, 0.12265580749511719, 0.12248262023925781, 0.12276019287109376, 0.12253388977050782, 0.12335411071777344, 0.12312371063232422, 0.12290866851806641, 0.12254515075683593, 0.12248780822753906, 0.122534912109375, 0.12261682891845703, 0.12288716888427734, 0.12256460571289063, 0.12276838684082031, 0.12297830200195313, 0.1227694091796875, 0.12250418853759766, 0.12257279968261718, 0.1226977310180664, 0.12254617309570312, 0.12250009918212891, 0.122787841796875, 0.122461181640625, 0.12268748474121094, 0.12251545715332031, 0.12266802978515624, 0.1225871353149414, 0.1227623062133789, 0.12271507263183594, 0.12253900909423829, 0.12249292755126953, 0.12263219451904298, 0.12244684600830077, 0.12256870269775391, 0.12278272247314453, 0.12261170959472656, 0.1223895034790039, 0.12428390502929687, 0.12278988647460938, 0.2565724182128906, 0.12282470703125, 0.12249190521240234, 0.12308480072021484, 0.12364595031738282, 0.1230540771484375, 0.12276640319824218, 0.12278163146972657, 0.12248473358154296, 0.12251136016845703, 0.12247654724121093, 0.12248678588867187, 0.12250009918212891, 0.12251955413818359, 0.12259839630126954, 0.12254617309570312, 0.12253593444824219, 0.1236684799194336, 0.12317388916015624, 0.12310221099853516, 0.12281139373779297, 0.12291379547119141, 0.12274483489990234, 0.12282572937011718, 0.12260765075683594, 0.12258812713623046, 0.12272946929931641, 0.12461772918701172, 0.12312989044189453, 0.12266185760498047, 0.12315033721923828, 0.12284620666503906, 0.1230540771484375, 0.122640380859375, 0.12250624084472657, 0.12265676879882813, 0.12255538940429687, 0.12278272247314453, 0.12277555084228516, 0.12254003143310546, 0.12262400054931641, 0.12318208312988281, 0.12256050872802735, 0.12256153869628907, 0.1224816665649414, 0.12287999725341797, 0.12276326751708984, 0.12258406066894531, 0.12270182037353515, 0.12269465637207032, 0.12290150451660156, 0.12344525146484375, 0.12252265930175782, 0.12313801574707031, 0.12258406066894531, 0.12265471649169922, 0.12261888122558594, 0.12246323394775391, 0.12254924774169922, 0.12281753540039063, 0.12254617309570312, 0.12263526153564454, 0.12243353271484375, 0.2562088928222656, 0.12256050872802735, 0.12245913696289062, 0.12246221160888672, 0.122461181640625, 0.12328652954101563, 0.12252774047851563, 0.12238028717041016, 0.12253183746337891, 0.12281037139892578, 0.12281651306152344, 0.12248678588867187, 0.12256358337402344, 0.1226792984008789, 0.12369407653808594, 0.12281958770751954, 0.12281549072265625, 0.12272742462158204, 0.12273356628417968, 0.12324864196777344, 0.12265676879882813, 0.12252365112304688, 0.12428800201416015, 0.12306227111816406, 0.12254208374023437, 0.12252979278564453, 0.12400844573974609, 0.12299775695800781, 0.12261376190185547, 0.12311654663085937, 0.12299980926513672, 0.12294143676757813, 0.12260044860839844, 0.12261682891845703, 0.12266700744628906, 0.1231677474975586, 0.12294656372070313, 0.12281651306152344, 0.12273766326904297, 0.12263731384277343, 0.12283084869384765, 0.12253900909423829, 0.12256768035888672, 0.12280937957763671, 0.12265058898925782, 0.12437811279296875, 0.12318822479248047, 0.12397875213623047, 0.12310733032226563, 0.12303462219238281, 0.12267314910888671, 0.12365312194824218, 0.12290662384033203, 0.12257689666748046, 0.12280012512207031, 0.12264447784423828, 0.12267110443115234, 0.1226455078125, 0.12263731384277343, 0.12309503936767578, 0.12316057586669922, 0.12300800323486329, 0.12256265258789062]",tokens/s,7.995929724812099,,,,,,,, -4bit-awq-gemm-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,internlm/internlm2-20b,internlm/internlm2-20b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemm,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch - benchmark_report = Benchmark.launch(benchmark_config) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch - report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch - raise ChildProcessError(response[""traceback""]) -ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target - report = worker(*worker_args) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run - backend: Backend = backend_factory(backend_config) - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 79, in __init__ - self.load_model_with_no_weights() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 255, in load_model_with_no_weights - self.load_model_from_pretrained() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 169, in load_model_from_pretrained - self.pretrained_model = self.automodel_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 558, in from_pretrained - return model_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3626, in from_pretrained - model = cls(config, *model_args, **model_kwargs) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/f363ea8a116b3ea829c7a068ca24bc9d3e668083/modeling_internlm2.py"", line 974, in __init__ - self.model = InternLM2Model(config) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/f363ea8a116b3ea829c7a068ca24bc9d3e668083/modeling_internlm2.py"", line 796, in __init__ - self.layers = nn.ModuleList([InternLM2DecoderLayer(config) for _ in range(config.num_hidden_layers)]) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/f363ea8a116b3ea829c7a068ca24bc9d3e668083/modeling_internlm2.py"", line 796, in - self.layers = nn.ModuleList([InternLM2DecoderLayer(config) for _ in range(config.num_hidden_layers)]) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/f363ea8a116b3ea829c7a068ca24bc9d3e668083/modeling_internlm2.py"", line 598, in __init__ - self.attention = INTERNLM2_ATTENTION_CLASSES[config.attn_implementation](config=config) -KeyError: 'sdpa' - -",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +4bit-awq-gemm-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,internlm/internlm2-20b,internlm/internlm2-20b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemm,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,,MB,3088.633856,14093.385728,0.0,13446.938624,13175.804928,s,10,15.872446411132811,1.5872446411132812,0.001977784195727725,1.5869342651367186,1.5899507080078126,1.5901222045898438,1.5902594018554688,"[1.58719189453125, 1.584764892578125, 1.5848123779296874, 1.5860927734375, 1.5866766357421875, 1.5853348388671875, 1.590293701171875, 1.58790966796875, 1.58991259765625, 1.58945703125]",tokens/s,161.28578630477753,kWh,1.8717543135086694e-05,1.0257253674026287e-05,9.112487845540085e-05,0.00012009967526451382,tokens/kWh,2131562.7992846123,MB,3091.492864,14093.385728,0.0,13446.938624,13300.586496,s,10,928.4395625000001,92.84395624999999,0.007570288473903254,92.84323828125,92.8531703125,92.8533390625,92.8534740625,"[92.8279765625, 92.8535078125, 92.8414140625, 92.8458828125, 92.8531328125, 92.853015625, 92.843109375, 92.8433671875, 92.8399140625, 92.8382421875]",tokens/s,0.6785579002079632,kWh,0.0010960241793427203,0.00060071625912522,0.005404193295573401,0.007100933734041341,tokens/kWh,8872.072654048685,,s,629,941.2515487060549,1.4964253556535052,0.1885841806933767,1.4736722412109375,1.4742849609375,1.4745136962890626,3.0601340332031253,"[1.4728734130859376, 1.473666015625, 1.473407958984375, 1.47373974609375, 1.4734244384765625, 1.47390771484375, 1.47325244140625, 1.4737960205078124, 1.4734847412109375, 1.4734468994140626, 1.473364013671875, 1.47363330078125, 1.4732943115234376, 1.47340087890625, 1.4733291015625, 1.47338134765625, 1.473206298828125, 1.473244140625, 1.4738125, 1.473511474609375, 1.4734613037109374, 1.4737889404296876, 1.473449951171875, 1.473580078125, 1.473765380859375, 1.473459228515625, 1.473560546875, 1.47323388671875, 1.474029541015625, 1.4734468994140626, 1.473263671875, 1.4734326171875, 1.47308642578125, 1.4729144287109375, 1.4731171875, 1.47327490234375, 1.473290283203125, 1.4730946044921875, 1.4730894775390626, 1.4733568115234374, 1.4730352783203124, 1.473344482421875, 1.4731263427734376, 1.4731683349609375, 1.4734151611328126, 1.4735421142578125, 1.473369140625, 1.4731141357421875, 1.4733404541015624, 1.47325439453125, 1.4731715087890624, 1.47344384765625, 1.4731990966796875, 1.473396728515625, 1.473218505859375, 1.4735196533203125, 1.4733302001953126, 1.47679638671875, 1.473838134765625, 1.4735543212890625, 1.4735584716796875, 1.473474609375, 3.058888671875, 1.4736353759765626, 1.4734478759765626, 1.47342333984375, 1.4732247314453124, 1.4741309814453125, 1.4735350341796876, 1.4742303466796876, 1.474051025390625, 1.4743828125, 1.4741749267578126, 1.4742200927734375, 1.4739588623046875, 1.4743316650390625, 1.4739302978515625, 1.4737408447265625, 1.474087890625, 1.47405517578125, 1.4743592529296874, 1.4743214111328125, 1.474566162109375, 1.473896484375, 1.47449853515625, 1.4739251708984376, 1.47448828125, 1.473300537109375, 1.4734755859375, 1.4742425537109376, 1.4745487060546876, 1.4744595947265624, 1.47428662109375, 1.4733526611328125, 1.4731519775390625, 1.47367529296875, 1.4740152587890625, 1.4736251220703125, 1.473764404296875, 1.4738944091796875, 1.4733157958984375, 1.473659912109375, 1.4736898193359376, 1.47346826171875, 1.4733157958984375, 1.4738472900390625, 1.4736722412109375, 1.4738759765625, 1.4739302978515625, 1.473797119140625, 1.4737530517578126, 1.47359130859375, 1.4737294921875, 1.4736865234375, 1.4733988037109376, 1.47375, 1.473734619140625, 1.4735994873046876, 1.4738953857421875, 1.47376123046875, 1.4736036376953126, 1.473574951171875, 1.474008056640625, 1.473680419921875, 1.473565673828125, 3.06031005859375, 1.4730966796875, 1.4736650390625, 1.4730096435546876, 1.473244140625, 1.4731519775390625, 1.4736844482421876, 1.4732667236328125, 1.4738687744140626, 1.4742548828125, 1.4740458984375, 1.473471435546875, 1.473302490234375, 1.4735523681640625, 1.473406982421875, 1.473101806640625, 1.47376953125, 1.4739149169921875, 1.4731151123046875, 1.4732298583984376, 1.4734039306640625, 1.4731314697265625, 1.473111083984375, 1.4737376708984375, 1.4738216552734376, 1.4737899169921875, 1.4738800048828125, 1.4739691162109374, 1.4734581298828124, 1.47338134765625, 1.4734776611328124, 1.47355029296875, 1.473333251953125, 1.4734776611328124, 1.4733946533203126, 1.4732882080078125, 1.473912841796875, 1.473892333984375, 1.474216064453125, 1.4738175048828126, 1.47407666015625, 1.4737275390625, 1.473375244140625, 1.4743492431640626, 1.4744185791015625, 1.4742845458984375, 1.4742108154296876, 1.474523193359375, 1.4741842041015625, 1.47373876953125, 1.4740675048828125, 1.474164794921875, 1.474018310546875, 1.473081298828125, 1.474177001953125, 1.4735738525390625, 1.47327587890625, 1.4737294921875, 1.4737100830078125, 1.4734151611328126, 1.4735892333984375, 1.47363330078125, 1.473797119140625, 3.0603828125, 1.4735103759765624, 1.4731171875, 1.4736036376953126, 1.4737879638671876, 1.4736036376953126, 1.47382470703125, 1.4734530029296875, 1.473491943359375, 1.4732840576171875, 1.4735810546875, 1.4739404296875, 1.4733045654296875, 1.4736578369140625, 1.4735308837890626, 1.4735565185546875, 1.4735206298828125, 1.47350830078125, 1.4735667724609376, 1.473364013671875, 1.4735780029296874, 1.473996826171875, 1.4732401123046874, 1.4733731689453125, 1.4740458984375, 1.473659912109375, 1.4739732666015626, 1.4740029296875, 1.47403369140625, 1.4737161865234376, 1.474050048828125, 1.474302978515625, 1.47325439453125, 1.473807373046875, 1.474555908203125, 1.474566162109375, 1.4731663818359375, 1.473481689453125, 1.4736650390625, 1.473850341796875, 1.4740213623046876, 1.4742425537109376, 1.4736527099609376, 1.4736947021484375, 1.473903564453125, 1.4739005126953124, 1.473833984375, 1.4741944580078126, 1.473732666015625, 1.4733824462890626, 1.473692626953125, 1.474039794921875, 1.4735994873046876, 1.4731766357421876, 1.4735155029296876, 1.473933349609375, 1.4738759765625, 1.473734619140625, 1.4740357666015624, 1.47370703125, 1.4738463134765625, 1.4742486572265625, 1.4741298828125, 3.061118896484375, 1.473859619140625, 1.4736097412109375, 1.4739149169921875, 1.4740848388671874, 1.474361328125, 1.4743634033203126, 1.4743951416015626, 1.473876953125, 1.4742210693359374, 1.4737510986328124, 1.473217529296875, 1.4732984619140626, 1.473310791015625, 1.473673095703125, 1.4737142333984374, 1.4734090576171874, 1.4737091064453125, 1.474240478515625, 1.4736937255859375, 1.4738544921875, 1.4743326416015625, 1.4735289306640624, 1.4736065673828125, 1.4740234375, 1.4740274658203125, 1.4741329345703125, 1.473870849609375, 1.4740899658203126, 1.473397705078125, 1.473323974609375, 1.473269775390625, 1.4733096923828124, 1.4737264404296875, 1.473578125, 1.47388916015625, 1.473797119140625, 1.474377685546875, 1.47427734375, 1.4738125, 1.473481689453125, 1.473944580078125, 1.473754150390625, 1.4732686767578125, 1.473934326171875, 1.4739149169921875, 1.4737080078125, 1.47405615234375, 1.4740255126953126, 1.4739588623046875, 1.4738177490234374, 1.4743765869140626, 1.474249755859375, 1.473743896484375, 1.4736650390625, 1.4745087890625, 1.473791015625, 1.473939453125, 1.4738421630859375, 1.4741165771484375, 1.4739957275390625, 1.4737919921875, 1.4739312744140625, 3.059681396484375, 1.4742149658203125, 1.4734141845703126, 1.4730894775390626, 1.4733916015625, 1.473896484375, 1.473638427734375, 1.4734898681640625, 1.4739404296875, 1.47435107421875, 1.47458349609375, 1.474293701171875, 1.4747064208984375, 1.4745426025390624, 1.47485693359375, 1.4739322509765624, 1.4736671142578126, 1.473491943359375, 1.4745078125, 1.474566162109375, 1.47476171875, 1.47451806640625, 1.4742261962890626, 1.4741903076171874, 1.47418115234375, 1.4742999267578125, 1.4743060302734374, 1.4741258544921876, 1.473733642578125, 1.4733138427734376, 1.472933837890625, 1.4730772705078126, 1.47359130859375, 1.473479736328125, 1.4740745849609376, 1.4737479248046874, 1.4733055419921874, 1.4737100830078125, 1.474017333984375, 1.4741851806640625, 1.47333935546875, 1.474234375, 1.4739200439453124, 1.473474609375, 1.47338037109375, 1.4740694580078124, 1.473732666015625, 1.4734735107421875, 1.4738309326171875, 1.4739005126953124, 1.473574951171875, 1.473660888671875, 1.47378076171875, 1.473206298828125, 1.4732236328125, 1.4738565673828126, 1.4739322509765624, 1.4736004638671876, 1.473474609375, 1.4738258056640625, 1.4735831298828126, 1.4737120361328124, 1.4735411376953125, 3.0642236328125, 1.4730126953125, 1.47343359375, 1.4729718017578124, 1.4735728759765625, 1.4736158447265626, 1.4739005126953124, 1.47361279296875, 1.473840087890625, 1.4736087646484375, 1.4734949951171874, 1.47426611328125, 1.47422314453125, 1.4739998779296875, 1.4739046630859376, 1.474639892578125, 1.474404296875, 1.474365478515625, 1.4745169677734375, 1.474093017578125, 1.4741513671875, 1.4736055908203125, 1.47390869140625, 1.47327490234375, 1.473606689453125, 1.473691650390625, 1.47382373046875, 1.473692626953125, 1.4730546875, 1.473733642578125, 1.4736373291015625, 1.47371923828125, 1.473250244140625, 1.473560546875, 1.4734674072265626, 1.473487060546875, 1.4741666259765625, 1.47329638671875, 1.473322021484375, 1.473585205078125, 1.4732359619140625, 1.473195068359375, 1.4736046142578125, 1.4733804931640626, 1.473373046875, 1.4742098388671876, 1.4742425537109376, 1.4739527587890624, 1.4736629638671874, 1.473869873046875, 1.47333935546875, 1.4739404296875, 1.4740797119140625, 1.4739158935546874, 1.47359130859375, 1.4735042724609375, 1.4738052978515626, 1.473397705078125, 1.4735718994140625, 1.473728515625, 1.4734254150390624, 1.473385498046875, 1.4733148193359376, 3.0616728515625, 1.4734571533203125, 1.4732861328125, 1.4731346435546875, 1.473143798828125, 1.4733568115234374, 1.4734940185546874, 1.4731695556640625, 1.4739505615234374, 1.473987548828125, 1.4735789794921874, 1.4733271484375, 1.4738790283203125, 1.4730577392578126, 1.473259521484375, 1.474188232421875, 1.474802734375, 1.4733424072265624, 1.473292236328125, 1.474060302734375, 1.4732933349609374, 1.4734765625, 1.47428662109375, 1.4740941162109376, 1.4739312744140625, 1.4739609375, 1.4747392578125, 1.4735155029296876, 1.47369775390625, 1.4734786376953124, 1.473407958984375, 1.473471435546875, 1.4732052001953124, 1.473833984375, 1.4732821044921875, 1.4737459716796875, 1.4735677490234376, 1.473176513671875, 1.4732420654296876, 1.473311767578125, 1.4738421630859375, 1.4735667724609376, 1.4740521240234374, 1.4739364013671874, 1.4734776611328124, 1.4733291015625, 1.4738514404296874, 1.473481689453125, 1.4737315673828124, 1.473935302734375, 1.4735574951171875, 1.473122314453125, 1.47314892578125, 1.473607666015625, 1.4732789306640626, 1.4734888916015625, 1.47401416015625, 1.47483544921875, 1.4746541748046875, 1.47443603515625, 1.4748702392578126, 1.4743603515625, 1.4745528564453125, 3.064785888671875, 1.4746173095703126, 1.47447607421875, 1.4746685791015626, 1.47447607421875, 1.4743643798828125, 1.4738125, 1.4731048583984374, 1.473474609375, 1.47338134765625, 1.473534912109375, 1.473302490234375, 1.473479736328125, 1.4740050048828126, 1.473471435546875, 1.47342333984375, 1.473850341796875, 1.4732401123046874, 1.4734808349609374, 1.4733658447265625, 1.4733629150390626, 1.473560546875, 1.473197021484375, 1.47348681640625, 1.473091552734375, 1.4735728759765625, 1.47310693359375, 1.473244140625, 1.4730096435546876, 1.4732452392578126, 1.473584228515625, 1.473293212890625, 1.4733302001953126, 1.4733189697265625, 1.473532958984375, 1.473896484375, 1.47376953125, 1.474218994140625, 1.473491943359375, 1.4740469970703125, 1.4737294921875, 1.4742374267578124, 1.4740889892578124, 1.473924072265625, 1.473764404296875, 1.473859619140625, 1.4736947021484375, 1.4739609375, 1.47409716796875, 1.47407763671875, 1.4737705078125, 1.4731356201171875, 1.4733404541015624, 1.473121337890625, 1.4731192626953125, 1.4733465576171876, 1.473375244140625, 1.4740101318359375, 1.4735616455078124, 1.47340087890625, 1.4733568115234374, 1.473385498046875, 1.4738780517578125, 3.063844970703125, 1.4731295166015625, 1.4731304931640625, 1.4736036376953126, 1.473776611328125, 1.47394970703125, 1.4742486572265625, 1.4738533935546876, 1.4738216552734376, 1.4736895751953125, 1.4738575439453125, 1.473681396484375, 1.473460205078125, 1.47378076171875, 1.4738780517578125, 1.4737530517578126, 1.473596435546875, 1.4732513427734375, 1.4733271484375, 1.47356982421875, 1.4737847900390626, 1.47350732421875, 1.473412109375, 1.47346533203125, 1.4734541015625, 1.473197021484375, 1.4740101318359375, 1.47386669921875, 1.4737366943359376, 1.4738145751953124, 1.4737264404296875, 1.47363330078125, 1.4732882080078125, 1.473638427734375, 1.4737899169921875, 1.473912841796875, 1.473796142578125, 1.4733824462890626, 1.473281005859375, 1.473376220703125, 1.473712158203125, 1.47343359375, 1.473532958984375, 1.4734254150390624, 1.4737110595703125, 1.473586181640625, 1.4736097412109375, 1.4738094482421875, 1.47342236328125, 1.473691650390625, 1.4737294921875, 1.4736414794921875, 1.4735841064453126, 1.4735462646484374, 1.4737427978515625, 1.473418212890625, 1.473576904296875, 1.473511474609375, 1.47342333984375, 1.4733507080078125, 1.47335986328125, 1.474207763671875, 1.47342138671875]",tokens/s,0.6682591926299519,,,,,,,, 4bit-awq-gemm-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,m,m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemm,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) @@ -2994,7 +2962,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4cb0-386a4bf7502b2760563b5a75;81b008be-9368-4add-82af-28412d2c727b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670f91e-42ebd26a7239d7f33d009346;8ae19849-bf36-4d6f-bac2-81a0fd2feed4) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -3064,7 +3032,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e559e-142cfdee20252ae550601e6e;d04e79c9-6a9d-4eb3-a00b-6fd76ef47593) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6671030b-486697cb5abc2499698bb325;75959508-0293-4286-a2a3-53966c76dac9) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -3134,7 +3102,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e54fa-395e7737130d4a4362eb547a;0711389a-2a2a-4db9-9cff-ed1ca9c25a44) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710258-7d6aa7f94ded993c3d6c7667;d68a0526-23bf-4ca2-a5f2-b281b6379e6b) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -3262,7 +3230,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e52c2-5e2935943afb868b6f2e34ee;476faedf-6460-470a-af2a-564aa9701014) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6671001b-74824eac10170d684f018d05;c86a8d15-0cf8-4d8a-aae7-2e87d8a2d824) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -3361,7 +3329,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e53bd-255ba6043fba403f73bfa1b7;588c3f8a-c7ae-42f6-88fe-32017c03dd30) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710116-38bc0166561773461112e063;2812c5b2-e781-4c49-b9f3-49ffb882e0b3) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -3649,7 +3617,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e43b7-30ca0a6a597c8c9c5bf193dd;5bdb9ed5-1588-457a-9ea1-940d86871133) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e720-23757b322512356d32ff87ed;9864e91b-40f0-47d0-b7bb-c85b53326538) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -3854,7 +3822,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5314-06ebfed5082e823b7e0bfe5b;c4a62e24-72e6-44fa-b1a5-1e2a931acfde) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710071-41c26ade494273a65944be59;c8d69176-a5c1-4b5d-926f-3efa34703169) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -3914,7 +3882,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e4360-271d48477910215e4422784e;a735e606-205f-44f4-be37-f774d05013ec) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e6ce-411b636265aca90f1c479e01;ff568d87-f33b-4fe4-a8e3-d1f8b8a58c20) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -4621,7 +4589,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e3e56-21d49eb3258a91772ab62c7c;2d5b2bae-0c13-4443-9feb-abceed23c5b9) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e191-06700e1f0344b2066e0d6bc1;72345123-2802-4c8a-9251-49ed95c319b0) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -4706,7 +4674,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e520a-7252e7f06e1e9f2b35c2785d;6d2dda6e-3d12-4a44-be56-f0d263c64c18) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670ff4f-2174e5a4667129a40a596a0a;56a32587-a141-4dc5-bfcb-f82f2c0e4ec9) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4952,7 +4920,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e505b-32fa484d1006671a74003ea9;bd1c8b63-29eb-4461-9d6c-b5ede6d9ec77) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fd47-620f24c50b4abdc66350646c;47b719c3-e4aa-4da1-9ead-8726d6df0a95) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5192,7 +5160,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e54b5-2b8b9c7043692c9b5f905345;4d0ec5f7-762a-4442-9188-f195755a0439) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6671020f-4019d24c1672f4024bc22e34;5db2df5d-8fd9-46db-9fb8-09485c723e26) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5291,7 +5259,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5007-43638d1c711936a909ef324f;d089090a-80e4-484f-89b0-e643540e98ba) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fce8-7ceb98334c9f423e14a8c6ec;66cfc2da-5fca-4bdb-a2d4-8c76c8acda2b) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5390,7 +5358,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e515f-2b0fa7d56669d1fd631811c1;cf58d4ae-b70a-4064-a486-7e5c957f2f91) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fe87-0e3e4e656e04cd90623c8e74;15a4581e-1556-4268-93d2-78064e7163d5) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5549,7 +5517,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e5948-6ebc2e732c0b08d61cc81996;25aa757b-8eeb-44ad-9669-1235114013d9) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667106bd-4a3d0ed74dde173746105505;0575c86e-91b5-402d-b453-1a084ad52a48) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -5634,7 +5602,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e51b5-564e6040422df46368a43173;808066eb-bf80-44c2-adbf-73bf21e67474) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fee1-7036601617b68e28073b25d9;0386bbc8-cb88-4500-8235-f5bfca9f5baa) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5758,7 +5726,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e430f-704bac141944a4353d05bc71;44680ae4-c03a-4945-bc7a-e70ac2384087) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e67a-1ce96c276aa2682608c866b9;c305379a-56b0-4908-ab92-d637b10c6b5d) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -5941,7 +5909,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e510b-5e3f8aec5792772e4c13e8bd;d7a44137-fb1b-4244-8af5-51a73a3a7ce7) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fe08-6934c8732eec168d5d7352cb;29f49513-a234-4496-9014-a7c22013d2ed) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -6104,7 +6072,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e440c-230af9ec6b2285a140376a67;b890a1a4-4b87-4189-bd47-def02a628327) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e774-7acc56f55830296e5f5c8fab;57d8b343-4bdc-42c6-830c-048578f9dac9) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -6267,7 +6235,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5410-60e7ce4833f702dc6c10723b;c3cc2d19-bdbd-43c3-a947-22432544d190) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710169-5de78cb4377f4499586c0dec;a0167c44-02e6-4ae8-ab0a-fc21ccae55d8) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -6401,7 +6369,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4d60-482ee4305f1497523b8095db;bdf216db-f2ac-4d12-9341-248a478ea875) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670f9d5-6719d7206e52cb0268be0e90;7cd438c6-595f-4625-a116-d5b50a3ad7c8) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -6866,17 +6834,19 @@ ChildProcessError: Traceback (most recent call last): self.pretrained_model = self.automodel_class.from_pretrained( File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 558, in from_pretrained return model_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3626, in from_pretrained - model = cls(config, *model_args, **model_kwargs) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/f363ea8a116b3ea829c7a068ca24bc9d3e668083/modeling_internlm2.py"", line 974, in __init__ - self.model = InternLM2Model(config) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/f363ea8a116b3ea829c7a068ca24bc9d3e668083/modeling_internlm2.py"", line 796, in __init__ - self.layers = nn.ModuleList([InternLM2DecoderLayer(config) for _ in range(config.num_hidden_layers)]) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/f363ea8a116b3ea829c7a068ca24bc9d3e668083/modeling_internlm2.py"", line 796, in - self.layers = nn.ModuleList([InternLM2DecoderLayer(config) for _ in range(config.num_hidden_layers)]) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/f363ea8a116b3ea829c7a068ca24bc9d3e668083/modeling_internlm2.py"", line 598, in __init__ - self.attention = INTERNLM2_ATTENTION_CLASSES[config.attn_implementation](config=config) -KeyError: 'sdpa' + File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3823, in from_pretrained + hf_quantizer.postprocess_model(model) + File ""/usr/local/lib/python3.10/dist-packages/transformers/quantizers/base.py"", line 195, in postprocess_model + return self._process_model_after_weight_loading(model, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/transformers/quantizers/quantizer_awq.py"", line 107, in _process_model_after_weight_loading + model = post_init_awq_exllama_modules(model, self.quantization_config.exllama_config) + File ""/usr/local/lib/python3.10/dist-packages/transformers/integrations/awq.py"", line 461, in post_init_awq_exllama_modules + model = exllama_post_init(model) + File ""/usr/local/lib/python3.10/dist-packages/awq/modules/linear/exllama.py"", line 144, in exllama_post_init + submodule.post_init() + File ""/usr/local/lib/python3.10/dist-packages/awq/modules/linear/exllama.py"", line 77, in post_init + self.q4 = exl_ext.make_q4( +NameError: name 'exl_ext' is not defined ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,1,64,1 4bit-awq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,m,m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,exllama,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -6918,7 +6888,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4cbe-505d252b2a82c3d2610b8802;fdb7175c-c9da-4f66-8236-82653a250576) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670f92b-1eb11a6019bfb72c15cc62df;0e2bea8b-ff5c-4134-9250-2206640a0f8b) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -6988,7 +6958,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e55ac-02ff0f762f5581e061005256;21f099c8-69bb-4abc-a1f3-5d9be496bf1e) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710319-1b042cf4702fab2820566220;73a3b211-3418-4ccb-a7e9-cc92f44b1237) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7058,7 +7028,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5508-2b249091718500905d7cd296;ca311b16-680d-4618-b1ac-e2d3e370e9f6) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710267-6fabb25e249797f62dee1ce9;5efe82b8-c13a-49cb-9e81-5a629f481ff9) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7186,7 +7156,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e52dd-13fe5cef7223f49224381c6d;84bc2cd8-61e9-49a7-aad6-557945fc416d) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710038-6c07c23d6ceabd32742a9c95;ff3ac86f-1611-472e-b6bf-932a83f90df1) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7291,7 +7261,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e53d8-333d2192741e6df65b00e467;ef52b333-0ec0-47b1-bbb7-7ce1477db702) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710132-14346c85190f2b0d64acf642;199ae124-21b7-4919-881d-a9f3724caa76) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7348,7 +7318,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpaphru3ca/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmprzwzsrpl/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,1,64,1 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -7591,7 +7561,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e43d3-5d69282e45fcf3db5b011b33;c030a0dc-244a-4de4-97a9-8d450c924738) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e73c-23212d103ee6f46d55eecdbc;03e021da-8b60-4f9f-a134-36941da8edbf) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -7816,7 +7786,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5331-1c14aed860d5b4f843cedaa8;ee6ad8bf-f863-46b5-b468-bcffe18c1e3d) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6671008d-0a00420e168ee78f450d5f03;cf2bdbee-a7ea-4db8-9e35-22f95410da99) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7876,7 +7846,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e437d-037dcd5560341ef0518615ea;ee42dce8-e807-479e-95f6-dc6c673b6f5e) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e6ea-7c8012867c753a9b53a63a00;ebe8e7b1-9ef8-42a1-9ae3-39f263397d07) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -8278,7 +8248,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpfk142j_w/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpcvqvw937/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,1,64,1 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -8625,7 +8595,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e3e75-5b7ede873be7537836313310;cfad55d1-a19f-4bfb-8b78-ffb3ad02c3b8) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e1b3-1d987a07796326152652a547;175985e8-98b6-46dc-a9a7-ba6c4434ec08) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -8710,7 +8680,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5225-304a38da0cf04fa472966b8b;353709aa-38da-4cd2-ab7d-d82e75749e85) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670ff74-0c3a7a474cf22ef076e608c1;4b621d1b-b284-444e-be75-acf78f02bfc9) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -8962,7 +8932,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5077-0c8e35e258c6283a58eae1b3;6d1d8d3f-5ac1-4f4e-a995-28ccc1c1cfad) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fd66-6ba1bc15694dbd77787af262;8ea35ca3-cdbf-4d82-9e8a-045229b00d68) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9096,7 +9066,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpycwf07_q/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpms8uxo4l/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,1,64,1 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -9202,7 +9172,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e54d0-4efd05a601ffb9c6062f6fe9;7fa60419-ea96-41ec-bb0f-6adff251ccdc) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6671022d-531a28d942bb599d1a2e4baf;d33883dc-bf55-4d9a-8f52-4f4acad669db) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9259,7 +9229,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpy9h0y78d/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpbiea8tvw/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,1,64,1 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,M,M,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -9301,7 +9271,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5023-0817a6d97fa59d017854457f;8c5b874c-0f68-4657-9717-f67fc8b889a6) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fd0a-27a3264e16a0528e446a158f;06c85555-9afa-4325-b669-33a530213159) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9406,7 +9376,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e517b-16c3966a2494983e79dfd1b3;9a574760-d7a6-4c50-96df-22bef75e90a7) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fea6-1f46fef75e49f7741caaad9a;4fad015f-ce7c-43c8-bab5-0c32d7f8c52b) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9498,7 +9468,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpugoxek9d/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp2gl40__1/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,1,64,1 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -9565,7 +9535,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e5964-58e963eb58897494571c614d;d531d837-44ee-4e53-b1f1-ed0d49c017a4) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667106d9-1f7bb87057ea2cd8600e97ed;92754f30-4b50-4025-956d-84cf3e546b01) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -9650,7 +9620,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e51d2-361776146a766e2a34a05e2e;476de83c-5e16-4b40-a7fd-3c1983b7318b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670ff01-2db920d4750a8587070feda8;284205e3-04e8-427a-ab71-04933007c45a) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9746,7 +9716,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e432a-4959ebd536dfbf3e13f1d79c;385fff8b-898a-48bf-93c9-fccf8a886f3a) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e695-5eae3b513004c711182aa3c4;23f4c04d-42af-4612-a526-b26909bbe0b5) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -9941,7 +9911,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5127-23bd53ae74fb8fa400e98f77;c454bd4d-0be2-4fbf-bfff-8941a25cf3d3) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fe28-59c42f347b1b55ce78e13d3e;b85ef84c-a78a-4603-a35d-4cdeb4619a00) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10116,7 +10086,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4427-3fa9d7f809d8282f1eb49cd5;f3d9a2c0-4779-4a49-9ef6-4633887ddb7f) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e791-54b4fa635e293ad007c1b31e;46a78b50-fdfc-4b0b-b41c-cdc3119a4a18) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10243,7 +10213,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpymjrne6x/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpa_g83la2/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,1,64,1 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,B,B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -10285,7 +10255,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e542b-197a6a193035509a5555f743;f318def6-b7fb-44c7-acc5-6e7f39fc50a0) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710184-0880748c3537228e3b6dac63;7bc543db-e67f-48c8-8086-7713f9a0ecf6) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10342,7 +10312,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpd6lbjnfv/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmppv1yf9c9/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,1,64,1 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,TencentARC/Mistral_Pro_8B_v0.1,TencentARC/Mistral_Pro_8B_v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -10419,7 +10389,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4d7e-0956a1c53e9a222717e3edcf;de436844-bac6-4652-b7fa-7dc52ff3ba9e) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670f9f1-33913a733f7b357475e09612;d6cc5142-7c29-43f9-8951-0addcf80ca91) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10511,7 +10481,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp_y_se4un/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpdqw4p8la/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,1,64,1 4bit-awq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,facebook/opt-350m,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -10947,7 +10917,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4cd9-32898e64517b539d41af2d42;54a7fc49-c558-4cbe-a477-f4feb6747570) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670f946-59ac66ef06499de05ad55706;556e8fea-6994-48d0-b61f-99dbd0b5055e) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11017,7 +10987,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e55c7-56a1996701299a2c79712c24;f9bddeb3-393f-41ba-a067-440519ec5f2c) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710335-15fecf5f57d8dcb31a34ffd6;7e266537-3c97-4748-b66c-05bbec5d51fd) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11087,7 +11057,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5523-05973ed04e2ff6b37e83f8a9;b2ef9a87-8354-46e5-8c5a-2bb6090f1375) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6671028b-6177ad5767644c923c48380b;83d6a4d7-b154-4412-a0eb-7801601c6c80) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11221,7 +11191,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5296-3efabbfc2fc39a8064493f5e;6751a8be-8c02-4b92-a8a8-ca9e85a54a02) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fff0-6f5ba6fe418612810672b8e9;6893b4a6-1845-4016-86ef-4174a21da81b) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11292,7 +11262,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5393-342b2c7210604e1c5895ab07;3b6c4bde-8d97-42ef-95b5-b82d62377b81) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667100ed-38368686610c7c885fe0df46;4bcc2d2b-528c-4e35-a677-71094f05e29f) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11476,7 +11446,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e438a-14a1088e384dcbbc028be28f;35bfcd92-e479-4afd-a3a6-0d824aefdde9) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e6f7-09558fbe05d0106606ce0158;2acc6d7b-a13f-48c6-81d0-c60bbf05aec8) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -11565,7 +11535,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e52eb-42e6cbf718e8878608aa09b4;2c8eff17-2140-458b-b7b7-f7fe1ea68f7f) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710046-3463d58c3c0dca374903eece;fa64a047-169b-426f-98e4-b7a55473b83d) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11625,7 +11595,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e4338-746ca09c37269fbd265ba877;347038b0-879c-438f-969d-fd5bfa71745c) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e6a4-662ee1b7229f481671a02d20;ec4028fe-e820-4cee-a826-edbe1eef278c) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -11909,7 +11879,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e3e2d-1e78efb478cc05ac33a83082;1a553d07-48a4-43cc-84f4-ceb871d26ffb) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e161-58c8959b585eafe9738a155e;be32811b-77cf-475b-bbe4-6ea07dff3366) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -11994,7 +11964,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e51e0-0ba84d4b6ec479bd5b2cb1bb;0d901ce7-9f13-43a6-bb26-3bf3a0f48140) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670ff16-6f11e2350789cd990f66bab7;f7c0e209-540f-4ef0-814f-2782a6f1691c) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12110,7 +12080,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5031-254ed7be2a39f7cf4a3a772f;6d2b4500-c6f0-4970-848a-bdbf3e39f84a) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fd1a-479f46e909a9693c649fac98;4145839c-bf73-4f8f-b95f-fe3ac0f2dfd0) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12254,7 +12224,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e548b-560efd2312fe7b1d2092c858;4c36e32d-54e1-4dd6-8151-2553b0a8fdcb) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667101e4-2605f5e904b44a6c78354296;95b1a5aa-1556-4ec7-b9bf-ca5d4184337d) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12325,7 +12295,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4fdc-7ff27be9056fbf21797a0944;6bffe524-c426-4dd1-9ac8-a9a5be9b817f) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fcbb-01f16c300c53ee51453ba2c0;bb195fbb-bbf1-4c16-890c-55e535fa8b33) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12396,7 +12366,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5135-0467e01d20aca62553d3a9a0;77bc6bd0-c6c1-449e-962a-2886bcd51609) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fe37-4eec62822bd1d6823b806690;8cdceea1-18f2-43f0-867a-eba30e7ec4eb) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12498,7 +12468,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e591e-6aa51a482f0e6dfd2e7a4fbc;9b84f52b-a66b-4be9-aa82-0e9e866e4822) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66710692-2274cdd4777ec9cb2ffaf810;a3f6e93e-4434-4f5c-9acc-ea2c9b320f33) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -12583,7 +12553,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5189-20fd729e43067b12609e6709;6fc2299c-070a-4ad6-8e50-7e5db54c753a) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670feb4-4cab84e636fda7a36c0a8df4;0e173d35-f85f-4513-848c-08c73d5d8218) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12645,7 +12615,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e42e5-737662ba05c72ac77706c47a;3d00de3b-f4d3-4c00-adef-589927cdba06) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e650-5c44ce50039cd8f90792f6f0;4e29dfce-e1de-474c-80bb-6367cfd5b477) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -12772,7 +12742,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e50e2-71bbabe102151ca24758b93d;169fbfd0-a6dc-4d02-89f0-6873acce89f2) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fdd8-46b8517c754301b62561a1ef;5cc6044c-6169-4dab-be0b-df7cb9c76e8f) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12845,7 +12815,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e43e1-1ff138a629ad77a46e609f48;a0276e37-80af-441a-94ef-d59da54303e0) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e74b-0c48b71c291266737d7ef417;19eac9b5-c62c-4b3a-aa0b-a26933b7f663) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12918,7 +12888,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e53e6-6861dd2e24b09c9d759e8e83;64f34151-d88b-415f-94d5-75983eafc6e4) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710140-72c8bf0542d2bb920765b39a;1a0e1210-a556-4476-800d-119ff008d80c) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -13054,7 +13024,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4d37-49940b3d6f66c7d154cb86fe;9bf01f5e-b361-4d68-97e2-64f2d8134bf4) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670f9a5-7000ba8d4bbf4bcf1c6ccc9f;294e1f0e-05f2-4482-bce9-22d6a8a801a0) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -13248,7 +13218,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4c95-7020879b381421d47a92e530;c6464d87-a1b0-4070-9ca9-1812e76049d3) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670f900-4475b3ca3fc4c4c55e6ecdc4;130bd415-7492-4205-84c4-3307115393d4) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -13318,7 +13288,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5583-5331eddf3e8e3f6a656395c7;cc2cde60-e512-4404-9a65-401831f7639e) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667102f0-1ab1b57a743cdaae3a1d263d;410685f7-6d70-4dce-ae5c-d79c395c2bc1) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -13388,7 +13358,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e54de-0a854b6e45c269c36843022f;c185d2f7-8dae-4536-ade5-b5505e6d5513) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6671023c-3f2fac127674d09d23ba57e0;8cc6d67d-ef06-4bf0-8845-dd7983ef1685) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -13488,7 +13458,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e52d6-3d5092bc19eab714532cd645;13cac4fd-dced-4c0c-bdc2-689797fe817e) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710031-03a90d1611aca6ad7a4b40e1;5726890c-f839-481f-bdd2-17bd19ee709e) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -13617,7 +13587,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e53d1-680313e61a1395cd71e93c1d;83778968-5c7e-49d1-a386-d8cbf0e64ba4) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6671012a-2d49903c0d4f6a7f1935df0c;4d1c21cb-f3db-4878-90f7-a0ba382353ff) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -13674,7 +13644,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpnobggdls/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpjjxcy53a/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -14003,7 +13973,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e43cd-1cfde75e7026632d67211069;2be23de6-6865-41c3-be1f-a5be10efef37) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e735-2bba20933a978da92b67874c;a24d56c6-66ba-419b-a0b5-c1a53b7f8538) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -14324,7 +14294,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e532a-2c481fea72ec62670c3bdc0e;288926d5-027b-469b-bb4c-0b9bc64b1d76) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710086-41d6b2176f779830035f692f;b5e95db8-f055-4017-8bd2-87f874d6c5d6) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -14384,7 +14354,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e4375-69da8da373ceb76f3b046271;5c89e8ee-5fb4-4838-8551-483210b00703) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e6e3-60c256514342b1a83240531f;9b123f80-4121-414f-b8d1-ed648222ba57) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -14906,7 +14876,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp3nh_z9qn/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpydlx8oda/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -15456,7 +15426,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e3e6e-2527c25972829aed5640da99;679e81b9-a3ef-4537-972e-b076c80129d6) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e1ab-7ed12f6d7883360c3f6df014;138c7a3a-f4a2-4c80-b4a0-107509ace7a9) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -15541,7 +15511,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e521e-034da3651f93012f1f8856c5;a1029e92-00c4-4b4e-9bef-b885989965bb) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670ff6c-077d704e3f4b6ca006ead21b;482315a3-d608-4e09-9e57-b4578860f834) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -15889,7 +15859,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5070-45c755543534b5e572e8fc5e;5e0e0da6-0ff9-48c3-a719-089c04ecf9f6) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fd5f-3c40912f06ddb1f5645655ee;1bca9017-5dc6-4fe6-acd3-9e68cf1bb4ff) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -16047,7 +16017,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp7ll2nqk2/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpccb2ftoa/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -16177,7 +16147,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e54c9-17f5d3a16a343c0f50d838ce;1fa8fce9-6b9e-41b1-97e3-5474d58fadb6) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710226-3030e96653b1575446f5f5b4;0ff60177-f585-47d3-ae88-57c8c5e77dd6) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -16234,7 +16204,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpdv2nx3g6/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp4sto5jc5/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,M,M,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -16276,7 +16246,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e501c-22e487082840c2b84572c8e3;8f709367-857d-4142-af1a-d39a9e21a6c0) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fd00-76d6a5a72787881e03a0a55f;c3328040-7d60-4477-8db1-1b3224708061) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -16405,7 +16375,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5174-05f80ef37c8db66f466ead78;bb106aa3-8038-42a8-ba1c-e3bb1987b7e4) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fe9f-128a954c7b80c85d113daccf;35a8aead-2b36-4d0d-a3e0-863bc8db138a) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -16521,7 +16491,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpxvr0fcut/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpp_9e5acl/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,01-ai/Yi-34B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -16612,7 +16582,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e595d-0cae55640d9cc40f05102f1d;59968847-24ea-419f-82c8-900050846285) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667106d2-3745237659a75aef7943d5bc;bda02f77-360d-49d0-9bf7-df27aa04af04) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -16697,7 +16667,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e51ca-060d3b6d109d00f3332ac1a9;15e2db9f-365c-4c89-a33d-ecf178a924f4) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fef8-204954de4847ccf34d6841da;3beda95d-0c30-421b-937b-d21a9eca9947) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -16817,7 +16787,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e4324-258063785707dde75dd14b4c;d697333d-6560-494d-a6e8-cb43487d945c) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e68e-61edd25d23af78a5053cf45d;44f26ed7-67b4-4e54-8ac5-a4cb75ebd2f4) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -17060,7 +17030,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5120-35e2e1ce4739efe22a91716f;e12ad655-92f9-4f41-b095-e131668d3096) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fe20-307ffabe34d3a9b529d3b7e4;b73cbc22-340f-4c34-b4eb-f45ece5cd594) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -17313,7 +17283,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4421-57244e631a97246128500d9b;5a227044-f240-4c7a-bf3d-26b025076eb1) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e78a-4ae7db96668068eb31c6854a;8339e63a-d06e-41d2-8179-f594480171e7) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -17488,7 +17458,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpmtjaoszw/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmps7pmci9v/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,B,B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -17530,7 +17500,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5424-1ed76dd369cb771d64fa529b;7b8174e7-cabe-44d3-90cd-abb241c1da44) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6671017d-6199141f04f4087c1992a5f4;4bcf717b-8f40-4fbe-9b26-36f7e0ddb880) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -17587,7 +17557,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpndfbyc5d/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpsaefxu24/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,TencentARC/Mistral_Pro_8B_v0.1,TencentARC/Mistral_Pro_8B_v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -17688,7 +17658,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4d77-06ea5cd037060f2f665c74c5;4ebe2b14-eb56-4972-9951-3aa49b099856) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670f9ea-47e13e9173db3c855211f6bb;6f66e196-9768-4a95-ab04-90d74433ea6b) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -17804,7 +17774,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpsydsuvyl/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpiwrw4af4/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,facebook/opt-350m,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemv,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -18356,25 +18326,25 @@ ChildProcessError: Traceback (most recent call last): return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/f363ea8a116b3ea829c7a068ca24bc9d3e668083/modeling_internlm2.py"", line 1047, in forward + File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/d431f9e4014c895663f30e8942d6e3c1adb6b562/modeling_internlm2.py"", line 1202, in forward outputs = self.model( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/f363ea8a116b3ea829c7a068ca24bc9d3e668083/modeling_internlm2.py"", line 932, in forward + File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/d431f9e4014c895663f30e8942d6e3c1adb6b562/modeling_internlm2.py"", line 1002, in forward layer_outputs = decoder_layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/f363ea8a116b3ea829c7a068ca24bc9d3e668083/modeling_internlm2.py"", line 639, in forward + File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/d431f9e4014c895663f30e8942d6e3c1adb6b562/modeling_internlm2.py"", line 736, in forward hidden_states, self_attn_weights, present_key_value = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/f363ea8a116b3ea829c7a068ca24bc9d3e668083/modeling_internlm2.py"", line 446, in forward + File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/d431f9e4014c895663f30e8942d6e3c1adb6b562/modeling_internlm2.py"", line 408, in forward qkv_states = self.wqkv(hidden_states) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) @@ -18426,7 +18396,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4cd2-7f0d52364662efb858d08ac3;2466369e-4cee-4b1f-b105-79ce041e6587) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670f93f-0f6cd3d105acc956508048fc;2a5c79be-f51c-4ff3-8a46-591bc6a94ac6) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -18496,7 +18466,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e55c1-5197c5213136510a238040a0;e5e422be-1793-4c27-a6dd-572c959c38ad) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6671032e-7a713b1b6600311a3e43acc6;43ff5122-4929-4362-8e29-6d3839f50d64) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -18566,7 +18536,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e551c-135ea7f403c41fb031bfb461;ee802eb8-842f-4da5-887c-0a82710dcc7a) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710283-328fdc597af9612409620e2d;d1dc6c69-924c-4749-8c09-43a08013181b) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -18724,7 +18694,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e52e4-46a1c9807553b2751dfd2160;f269c3f0-fad8-42fb-a4eb-9737bdc0444b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6671003f-3e459de76e38d3be433d6aeb;74232d1f-3ac7-4db7-ac75-3bd3f1fd399b) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -18829,7 +18799,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e53df-5dc26ee1660494f4725f3237;cf59816b-89d8-4b58-a626-b16571a440e1) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710139-053004431084d4d119be619b;1097fe3b-7412-4276-a478-92abe16839d4) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -18886,7 +18856,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpxp5xnrna/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp74xa8375/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,2,64,1 4bit-awq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -19127,7 +19097,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e43da-34048b6761e91daa01f3d3e5;1d19fe01-91fa-4e16-815f-28c0dba2dd11) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e744-34dbcf8f734c1e8f0c3abf41;4e079611-63c7-436d-9143-64e123ed8ea9) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -19350,7 +19320,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5338-2cb857ad1514178b32c57029;c3c11dd0-fe3e-456b-9db4-eb42981fa285) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710093-284c0d4b1ad0c1695c9a6124;1440b143-366c-420e-81a7-44189f6a51a9) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -19410,7 +19380,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e4384-5af7cc150d9456cd11a6a82e;eb635b80-22eb-4089-be9f-668f5f9a3ca0) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e6f1-2737e0a7455554392c365da8;7e9c5e78-2658-49be-8b78-fbcca9326fe0) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -19814,7 +19784,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpa_dnl6rw/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmppr2x8zex/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,2,64,1 4bit-awq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -20161,7 +20131,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e3e7c-2c969882564c33bd64972e36;e7b2d4df-5c10-42a0-8413-aa44ca687124) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e1bb-4a99ab765385a18679d8cc1a;6562b007-976e-42c6-9490-7f8cf1dbd4e1) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -20246,7 +20216,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e522c-1625e0a2543e1ae76843da3f;05ce5f25-511e-4fa4-b109-17f7eed7eaf5) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670ff7c-409cd71f2f2e499b48446b21;5dce7573-eb73-4a7e-b2ab-75eacd9a241c) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -20498,7 +20468,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e507e-5076ba627b720c53772906ca;b687862a-ee4c-4c86-8428-3a816b964d93) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fd71-070c4b094b2ec40013663636;3ff8ff3b-c223-4358-b936-96a74ff0693a) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -20746,7 +20716,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e54d7-62ace1f11e306ef22914729f;66c4590e-94f6-4399-8e69-e5ee45cddfb3) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710234-2a5f4aea61c01bbc1e18d2cf;6b121ee2-93a8-40c1-8860-786bab674c50) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -20803,7 +20773,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpia81_gz3/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpzgc6ng2t/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,2,64,1 4bit-awq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,M,M,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -20845,7 +20815,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e502a-37a8845321b747c65209e9b4;a07348cd-e884-4701-bdf0-24028194caf1) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fd11-273c9ce809179f8d15f28822;e0e334a4-0d2d-4f52-a509-995906db00d5) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -20950,7 +20920,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5182-1dc8f3567bb1213d2bd99747;8d419f11-fe41-4bcd-82ac-ca959117c622) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fead-6e55a165471b50854722c8bc;e15ad000-976d-45a9-b1a4-68a60afbbd0d) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -21113,7 +21083,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e596a-49dfa34b2bd515226c4ac221;09c54195-d798-40fb-80c7-2fc1b1378233) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667106e0-4a6b29055569da4833e98ce6;01bb7d6e-8208-4023-ba92-5ecc45017611) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -21198,7 +21168,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e51d9-58e31c3c7c0c1e6e2bc991b3;c0294ab3-e8d3-4c7c-b2d2-bec9fa6bceee) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670ff0b-5ef1ff9255ee97557effec7d;a2354076-7ba3-4a33-a0df-dae6ef55ddb5) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -21294,7 +21264,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e4331-22bec2cb332949473e6ae244;89118fb9-30ae-4615-835f-58a462318359) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e69c-11488d9b293a29d36f22c0aa;21bd5466-5ca5-4d02-9c53-8d19d5925c72) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -21489,7 +21459,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e512e-17a978463fce98ef34c584b2;411ab227-9f0f-47cd-b90b-d2131519959c) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fe30-41ead6c96c5016207480444a;71db91f1-7700-42d3-8b93-5a90c6545858) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -21664,7 +21634,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e442e-085a2e5153d0425248abe0b4;07748ad6-f81d-46e8-8118-208f14ef4bf5) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e798-06cddfa65e6cc6ad05ac7882;e8d8b929-9ece-47a6-8731-4f5f54f45450) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -21791,7 +21761,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp2150iz0c/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpl_cybicq/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,2,64,1 4bit-awq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,B,B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -21833,7 +21803,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5432-1adb8a4d6b813a7169f3962a;0182457f-0569-43a2-a9db-b101f3b6b1fc) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6671018c-5b10d40829bdce8c50c19a9e;06a16946-be88-4e68-81cb-e74b9052035b) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -21971,7 +21941,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4d85-628797975c63d0da30158ba8;6871e7d0-e105-470e-b313-eaebdefc60a6) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670f9f9-7f7dd2f0694ae7012fdf99c2;2c676ad6-773e-4e60-acf4-e29fe23c7d42) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -22063,7 +22033,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpvtqs9z15/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpn7pusk0n/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,2,64,1 4bit-awq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,facebook/opt-350m,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,exllama,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -22501,7 +22471,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4cdf-5aef56127fed928a5da3cf18;9ab623bb-c57a-4388-a424-0ecc3fec7ce6) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670f94d-68dbd118370fa62a345a35a0;bf9749df-eacc-4d46-a6b9-c6dfcaeb21c7) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -22571,7 +22541,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e55cf-53d8c54f7129a06633a82937;756920b0-108b-4a0a-8692-fc31b387c73b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6671033c-30f611f804353e157196430c;04b9d5ea-f054-453a-8b80-aa743d098a46) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -22641,7 +22611,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e552a-3b535e746f15f4417249aad7;e7deba9a-a585-48aa-85ad-3b7d16be74dd) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710292-2668420944182b610d19ebde;4386843f-fc0e-44e0-adf9-a164ad4a0a73) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -22775,7 +22745,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e52a5-1408ac2020e5863238438b2e;8d4b62b6-e9d2-49dc-bc83-e7ac09141789) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670ffff-32bcb44306c2f2ad2ace483b;f43934a6-4259-4d45-95ef-ed7ff270e1ea) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -22880,7 +22850,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e53a1-34b6e14f0b86075c6398c9c9;3f91cda0-8d78-4310-9dbc-4894da15f5cd) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667100fa-6720ff407482f1c40c920628;f069a90b-7f7c-4293-ba70-8ef0e378f59e) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -23186,7 +23156,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e4398-3743bda470fd22e34a664984;5cbf628b-e0e0-4a4c-b3a0-20acace37324) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e705-0ce439e445019307264154ed;38fe31ee-a32a-41c3-b542-153fa89caf9f) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -23411,7 +23381,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e52f8-6487b19b5be703e239bcbc30;5c4556ae-a715-4124-bbc5-0f03a8004480) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710055-59b07253198053a86ed33c80;3c72b05a-7c4c-4ede-b530-85b14085ee19) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -23471,7 +23441,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e4345-1d06edaf0d28cff85e4f6766;ea77b0ca-de22-48e3-97f4-959cb2177075) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e6b2-18e07881362c65722cb7a82c;e071f1a7-85a5-4e1d-92d4-827596609475) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -24226,7 +24196,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e3e3a-789f13a54811e1c14289eccc;4d43f16c-ed48-4764-bfe7-18b05a89611c) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e170-3850658718d72f6f0a06599c;025f1101-9522-4722-9ec2-555385e1368e) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -24311,7 +24281,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e51ee-2594d0874c2c766a56ecb034;0b432e07-5e66-4c0b-a11f-be3b5716ff79) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670ff27-26f09e3b6edc96ef796ea231;10dd21d8-8f7e-4abf-8e45-d5f5bd94e82e) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -24563,7 +24533,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e503f-347a00242edb348336dc498a;f697d4a2-2e04-4b6c-9433-c6922d90b963) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fd29-0387111c17da0b4f53d93f5f;5122bc6c-90cb-4884-9972-8c05d342e9b8) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -24809,7 +24779,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5499-606ce14f0fccedaf4d8c0d40;c36a130b-aec1-4445-b98e-cf780d5e6131) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667101f2-1724788e1a5c30d7495c7863;ac09023a-1752-4b54-ab5e-ae1275b658fa) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -24914,7 +24884,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4fea-60d42f1855057b2c7b90b3ad;2868eaa9-ae8a-4317-8372-2630ba5ac60b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fccb-1f28b176171be66c0776943e;5db8af7a-971a-40fa-8bcb-71be21a0a986) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -25019,7 +24989,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5143-164340b177b5e5cb2dd381be;4aa14db3-2e7b-47e1-a3b6-b20d29ee6d9a) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fe4d-702bb14b0495b6595641a6aa;0928f644-0809-49d3-9714-4b860fe1eef8) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -25184,7 +25154,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e592c-04443c7a13885f1520cab448;ca3db71f-a8e4-46ce-a7cd-77d64dcfd063) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667106a1-19f1fbd23304567e4e09e81c;879b8c05-723e-4e13-a417-ac4550876d57) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -25269,7 +25239,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5197-2aec114032c4ce4b175cd32f;e7393f95-f8c8-4abc-9c56-0a8531d87502) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fec3-374b8444476950424b48364b;9aac6907-d591-471d-98c6-b21a7728c64d) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -25365,7 +25335,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e42f3-64c4ee892f00c34d01fcb716;b13f3411-e798-4546-bbb4-fe3b91210e80) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e65e-4b8bd60d3e043b2c7905b6fd;4a31a7cb-2808-43d2-aa6d-70d8490f0fb1) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -25560,7 +25530,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e50ef-6eae7b55131dc51b400139e6;3c11d3a9-ec32-46f9-b9b6-76650cbcc549) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fde7-558d548f6177135e6b8cd2b4;1199e70f-2444-4ed6-83b4-0260fbbd2a93) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -25735,7 +25705,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e43ef-72369df369d55a002566159b;5566fe3e-64c9-47ed-b244-1ae6aae9365c) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e758-49f304bf2cfb9a521d680db1;1766e0eb-883e-4b54-b2a2-031ae83fe9a6) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -25910,7 +25880,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e53f4-30235f671cd67ff834b05379;a922d8b8-a73e-4781-b46d-570022bea879) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6671014e-1f3704155ec00d9b67343e36;01beb1c7-f417-438d-b6a2-6773b3372b5d) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -26050,7 +26020,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4d45-4346256531207595787a2455;0453bd28-cb5d-4c04-9f0f-c99050597ba0) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670f9b3-2d2ae59a34f1fc507c8a197c;67252ed2-eee6-47fa-8d84-d099b4b39e5d) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -26584,7 +26554,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4ca2-762bbaf1693b3ed2033fc2e9;b42f6906-e09c-42ef-be2c-ef0f8dc36a13) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670f90e-63c9027f1de3d877265bdd66;5bb97f4a-95eb-4937-87d1-6af3385c620a) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -26654,7 +26624,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5590-23f038a94690ff9b06a93ed9;70124663-4737-4b71-9c29-ad61e623656f) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667102fd-2a99d9b87196af28046eb82c;17d4fb46-2995-43e1-bd9d-8bbab8eb92c5) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -26724,7 +26694,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e54ec-438c6c407c3b32a250d68304;ea33c7cd-9b90-4c92-ad17-d4235611f1e2) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6671024a-467ed84407521ccb6e00442e;4ff36e55-25fb-4615-bd02-0f158872a57c) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -26858,7 +26828,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e529e-35bb3bce5a1c94cb5074d25f;1c6cf0d5-d9c0-411f-890e-3c8655dd802c) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fff8-5558f4be4ce697d213170beb;5275f33e-06ba-4bba-bb2e-533beb271345) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -26987,7 +26957,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e539a-0e25a53e4d3e53e271406377;0eed3c98-dacf-4909-a639-b3b27296dbd9) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667100f4-4c3f4e6c3961a7c647866a70;3b88aa42-4a98-4471-b6c3-a25d5ea0530a) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -27403,7 +27373,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e4391-62088da006effec639bfab52;27c4f517-feeb-4ae2-aa07-ecba15acdace) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e6fe-51a16200550daa573dfafa59;0d82d240-54ca-4515-ae2e-b521e2a31321) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -27724,7 +27694,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e52f1-61dcd3212fd154226174ef4e;ce313471-40da-4a39-be02-3478297a5920) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6671004e-74f205cd015e9d7004e1a066;5abb1044-aa24-4977-9569-e90710c9f015) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -27784,7 +27754,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e433e-2490d7ba61e4caaa64cafef9;b3b967ee-29d9-4605-acba-bec3189af598) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e6ab-2cc921da546863316cb2014b;35ffc8c5-4950-48fa-8131-588cbb0297c9) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -28886,7 +28856,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e3e34-475b405b1dad04633aebcdf4;781ae4bb-43f7-4582-8573-36213c7dbfd7) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e168-78b0227c6c6276e56e5e7e61;41898845-62ce-40bb-b03f-74062b24907f) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -28971,7 +28941,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e51e7-023131c305cfb9f81623de7a;cfa4b825-329c-467b-aa5e-23404dc8c936) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670ff1e-6d966dc74da90c844cb103e3;2f2358da-6a2b-4d96-82c4-3e7001e28683) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -29319,7 +29289,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5038-7d01a14610f1722a1d509600;685d1908-1f24-47f8-83e8-5d37757a24e5) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fd22-21447d7709f416736d45379e;7e1f094d-11cc-41a6-858f-80d1ca22e29b) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -29637,7 +29607,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5492-11c17722771a703b1fa351d1;763bef09-4b05-4198-821b-5a5d89280bbb) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667101eb-6751eddb4f4303a6025a25bc;5dc726dd-a03d-42da-974c-43df63b3e402) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -29766,7 +29736,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4fe4-781b9ab62e5b95645d8bd218;754aab59-6120-4381-a4fb-076e8a426fe0) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fcc2-14d2b8f226d59e2970305fc8;3f98aafe-b71e-4069-959d-f6c42664edb0) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -29895,7 +29865,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e513c-1f6c1bcc34e230325afbb04d;97ebb48f-9e2c-4969-a1f5-da1e8ac2e7de) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fe40-6af6f9a239ac119b0a2a9484;07fe9edd-6d2b-4e75-b083-cdccd6f08cf5) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -30113,7 +30083,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e5925-64b9dc341b22d1ee71f1e009;228a92f9-eef3-4dd0-ae10-f03544e2aa11) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6671069a-3ed1c1cd581ad5cb6dfae840;45dc0557-328e-4b50-9272-2f346dfc8900) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -30198,7 +30168,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5190-56a3f875624c7a1e286d4bcf;10423acd-0876-4b7c-af4c-3fc498036d65) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670febc-2558f6025d512ddd040210c1;8975db56-63cd-4560-b0f0-0d28031e9d56) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -30318,7 +30288,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e42ec-3691ccbc439acc075f54f583;6eb6cca5-4030-4bb7-9f2d-75f0b78c3ea1) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e657-1428525e538d72193b2e2bd9;8fe8f3f0-7fdd-468e-97b0-afa65bf18ae1) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -30561,7 +30531,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e50e9-32e621df69c15593439efb85;5c9bc7d5-6ac9-45b6-82da-c1b69d4908e9) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fde0-2c6ba22f2799e5972ebc60a4;61a7d49e-2776-46f9-b162-ee35bdfce474) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -30814,7 +30784,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e43e8-08fd103302f3fe3571ecbd6e;4f8b0e88-ba87-4034-994b-a12049a82e3c) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e752-7cbab47e30abc24044da67d3;bea452c6-6dd6-4e5d-9ee8-0debed705c58) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -31061,7 +31031,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e53ed-51ed6ab964aa94d017e682c7;15460d5a-7d3d-49f1-a3ee-2d7208872bb9) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710147-575e4fad5af7196f0f997215;d8885843-2e81-48f0-8864-7d49ef3a0edc) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -31249,7 +31219,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4d3e-510bf36b47d0907623eccc81;d0123cac-c983-4590-acad-8e4ee399f057) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670f9ac-575e26ed125d4ac105881141;3e7d623b-aa9c-48b4-80e1-bdbae58cac3d) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -31947,25 +31917,25 @@ ChildProcessError: Traceback (most recent call last): return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/f363ea8a116b3ea829c7a068ca24bc9d3e668083/modeling_internlm2.py"", line 1047, in forward + File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/d431f9e4014c895663f30e8942d6e3c1adb6b562/modeling_internlm2.py"", line 1202, in forward outputs = self.model( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/f363ea8a116b3ea829c7a068ca24bc9d3e668083/modeling_internlm2.py"", line 932, in forward + File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/d431f9e4014c895663f30e8942d6e3c1adb6b562/modeling_internlm2.py"", line 1002, in forward layer_outputs = decoder_layer( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/f363ea8a116b3ea829c7a068ca24bc9d3e668083/modeling_internlm2.py"", line 639, in forward + File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/d431f9e4014c895663f30e8942d6e3c1adb6b562/modeling_internlm2.py"", line 736, in forward hidden_states, self_attn_weights, present_key_value = self.attention( File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl return forward_call(*args, **kwargs) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/f363ea8a116b3ea829c7a068ca24bc9d3e668083/modeling_internlm2.py"", line 344, in forward + File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/d431f9e4014c895663f30e8942d6e3c1adb6b562/modeling_internlm2.py"", line 306, in forward qkv_states = self.wqkv(hidden_states) File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl return self._call_impl(*args, **kwargs) @@ -32017,7 +31987,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4c9b-296a40957aaeac240f64c3dc;5724d6af-ec83-4056-b774-c5444f354626) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670f907-3b130f8a7fbfe4c46a46f6f1;a2e0c816-2bc9-4f64-9927-ee20928d604a) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -32087,7 +32057,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5589-1c73c62f44d3c37a2e714e8b;bdf94ed2-2a62-41c1-9ea0-446f4c8fdbfd) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667102f7-6cd1687c16844c4c1e6db7d8;3fd6d2f9-bcab-4b18-98b9-e4e9712a5ba1) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -32157,7 +32127,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e54e5-7f9ca2fd7f13097e35150bf0;8f9d860e-2ae1-4657-abe2-48c52c38a959) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710243-307b582e4f503d8d0106157b;591a88a5-8f96-4c1d-9cff-50220e1c755c) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -32315,7 +32285,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e52cf-4fe7b41c24dea28d43ce524e;6e8d5fb7-1d85-4627-b6ec-0d2d0116450f) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6671002a-60b89a7517c56efd19717ba1;b13f421b-48ae-46db-a655-c82587d76fa9) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -32386,7 +32356,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e53ca-3ad699676ba450283b5f47d1;4abfa1da-ae9d-412c-afb2-974d1b528965) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710124-5d76db010b3ea28e6dd44d23;627d455e-39fb-4084-89a0-24ed7e21c981) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -32443,7 +32413,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpagror65d/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpma3dxokt/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,EleutherAI/pythia-70m,EleutherAI/pythia-70m,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.40.2,,0.30.1,,,,1.19.2,,,,0.11.1,,,MB,1279.725568,872.93952,0.0,226.492416,184.397824,s,12,0.17399129676818847,0.014499274730682373,0.0004405777377392672,0.014355583667755127,0.01475712022781372,0.015286664152145383,0.01576770728111267,"[0.01588796806335449, 0.014419008255004882, 0.014310144424438476, 0.01430396842956543, 0.014377471923828124, 0.014377087593078612, 0.01433407974243164, 0.014250847816467285, 0.014390144348144531, 0.01479468822479248, 0.014266719818115234, 0.014279168128967285]",tokens/s,17656.055544507362,kWh,1.7155437431226257e-07,9.399407219055296e-08,3.2474264005518264e-07,5.902910865579981e-07,tokens/kWh,433684339.52262825,MB,1280.053248,872.93952,0.0,226.492416,197.932544,s,12,10.420531616210937,0.8683776346842448,0.009197306451406238,0.8653596801757812,0.8716784362792969,0.8835345275878906,0.8948784387207032,"[0.8977144165039063, 0.866240966796875, 0.8642362670898438, 0.8648695068359376, 0.8719328002929687, 0.8635298461914063, 0.862948974609375, 0.866379638671875, 0.86938916015625, 0.865849853515625, 0.8638052368164062, 0.8636349487304688]",tokens/s,72.54908174012076,kWh,1.0699516429705703e-05,5.862862695252688e-06,1.9223555826850592e-05,3.578593495180898e-05,tokens/kWh,1760468.186309475,,s,755,10.563054601669307,0.013990800796912995,0.001781085498877292,0.013702143669128418,0.014027775764465332,0.014339276885986327,0.028389335021972658,"[0.015265791893005372, 0.014698495864868164, 0.014765055656433105, 0.014334976196289062, 0.014128128051757812, 0.014334976196289062, 0.014543871879577636, 0.014234623908996581, 0.0144650239944458, 0.01454694366455078, 0.014418944358825684, 0.01439027214050293, 0.01425100803375244, 0.014261247634887696, 0.014215167999267577, 0.014323712348937988, 0.014708736419677734, 0.014633983612060546, 0.01434931182861328, 0.01467801570892334, 0.014370816230773926, 0.014229503631591797, 0.014525440216064453, 0.014231552124023437, 0.014211071968078613, 0.014244864463806153, 0.014368767738342286, 0.014796799659729003, 0.014531583786010742, 0.014297087669372559, 0.014436384201049805, 0.014063584327697753, 0.014220288276672363, 0.013916159629821777, 0.014133248329162598, 0.014468095779418945, 0.014696479797363281, 0.014910431861877441, 0.014256128311157227, 0.014102527618408203, 0.01407590389251709, 0.014150655746459961, 0.014071807861328126, 0.013922304153442382, 0.013725728034973144, 0.01371132755279541, 0.013733887672424316, 0.013731840133666993, 0.014045184135437011, 0.014156800270080566, 0.014639103889465332, 0.014141440391540527, 0.013728863716125488, 0.013714336395263671, 0.013676544189453126, 0.013729791641235351, 0.013732864379882812, 0.013691904067993164, 0.013711359977722168, 0.013743103981018067, 0.013743103981018067, 0.013660223960876464, 0.0284968318939209, 0.013691904067993164, 0.01368883228302002, 0.013686783790588379, 0.013670399665832519, 0.01399295997619629, 0.013728768348693847, 0.013667327880859375, 0.013691904067993164, 0.01365503978729248, 0.013710335731506347, 0.013666303634643554, 0.013699071884155273, 0.013669376373291015, 0.013689855575561523, 0.013734911918640137, 0.013744128227233888, 0.01374617576599121, 0.01366323184967041, 0.013693951606750488, 0.013799424171447755, 0.013740032196044923, 0.013728768348693847, 0.013709312438964843, 0.013709312438964843, 0.01379532814025879, 0.013731840133666993, 0.013713408470153808, 0.013707263946533203, 0.013776896476745605, 0.01368883228302002, 0.013706239700317382, 0.013716480255126954, 0.013789183616638183, 0.014757887840270996, 0.013946911811828614, 0.013715423583984376, 0.013734911918640137, 0.013693951606750488, 0.01369600009918213, 0.013839360237121581, 0.013735936164855958, 0.013712384223937989, 0.013700096130371094, 0.013719552040100098, 0.013707263946533203, 0.013677568435668945, 0.013913087844848633, 0.013707263946533203, 0.013698047637939453, 0.013667360305786132, 0.013723615646362305, 0.013713408470153808, 0.013699071884155273, 0.013731840133666993, 0.013686783790588379, 0.013784064292907714, 0.01374828815460205, 0.013744064331054687, 0.013783040046691895, 0.013717503547668456, 0.013740032196044923, 0.013710335731506347, 0.028412927627563478, 0.013706239700317382, 0.013686783790588379, 0.013731840133666993, 0.013669376373291015, 0.013661184310913087, 0.013645824432373046, 0.013658111572265624, 0.013669376373291015, 0.013628416061401367, 0.013694975852966309, 0.013636608123779297, 0.013705216407775878, 0.013678591728210449, 0.013858816146850587, 0.01386086368560791, 0.01368883228302002, 0.013693951606750488, 0.013660160064697266, 0.01374518394470215, 0.013757408142089844, 0.013689855575561523, 0.013717503547668456, 0.01369600009918213, 0.013725695610046386, 0.013678591728210449, 0.013732895851135254, 0.013722592353820801, 0.0136878080368042, 0.013707263946533203, 0.013698047637939453, 0.013684736251831055, 0.013669376373291015, 0.013684736251831055, 0.01375641632080078, 0.013708288192749024, 0.013765631675720215, 0.013703167915344238, 0.013721599578857421, 0.013702143669128418, 0.013694975852966309, 0.013763584136962891, 0.01369600009918213, 0.013714431762695312, 0.013794303894042969, 0.013682687759399414, 0.013697024345397948, 0.013739007949829102, 0.013708288192749024, 0.013678591728210449, 0.013742079734802246, 0.013710335731506347, 0.013716480255126954, 0.013678591728210449, 0.013706239700317382, 0.013776896476745605, 0.013827072143554688, 0.013702143669128418, 0.013737983703613281, 0.013683712005615235, 0.013726719856262207, 0.013710335731506347, 0.013788160324096679, 0.028406784057617186, 0.013711359977722168, 0.013690879821777344, 0.013694975852966309, 0.013714431762695312, 0.013697024345397948, 0.013701120376586913, 0.01367142391204834, 0.013676544189453126, 0.013768704414367675, 0.01366528034210205, 0.013711423873901367, 0.013736895561218262, 0.013741056442260742, 0.01368064022064209, 0.013705216407775878, 0.013719552040100098, 0.013714431762695312, 0.013728768348693847, 0.013700096130371094, 0.01368883228302002, 0.013689855575561523, 0.013683712005615235, 0.013835264205932616, 0.013717503547668456, 0.013843520164489746, 0.013794239997863769, 0.013710335731506347, 0.013718527793884277, 0.013725695610046386, 0.01368064022064209, 0.013674495697021484, 0.013712384223937989, 0.013718527793884277, 0.013752320289611816, 0.013718527793884277, 0.013736960411071777, 0.01367347240447998, 0.013757439613342285, 0.01370419216156006, 0.013725695610046386, 0.013716480255126954, 0.013707263946533203, 0.013740032196044923, 0.013692928314208984, 0.013726719856262207, 0.013685759544372558, 0.013721664428710938, 0.013703104019165038, 0.01367142391204834, 0.013768704414367675, 0.01369600009918213, 0.013686783790588379, 0.013691904067993164, 0.01368883228302002, 0.013734911918640137, 0.013686783790588379, 0.01375334358215332, 0.013711359977722168, 0.013979647636413574, 0.013771776199340821, 0.013735936164855958, 0.013831232070922852, 0.028631999969482423, 0.013729824066162109, 0.013675488471984864, 0.01368172836303711, 0.013717439651489258, 0.013657088279724122, 0.01366528034210205, 0.013697024345397948, 0.013677568435668945, 0.01365503978729248, 0.013760512351989745, 0.013660160064697266, 0.01366220760345459, 0.013717503547668456, 0.013717503547668456, 0.013676544189453126, 0.013710335731506347, 0.013657088279724122, 0.013947903633117676, 0.013619199752807617, 0.013640704154968262, 0.01366528034210205, 0.013693951606750488, 0.01365401554107666, 0.013627391815185547, 0.013701120376586913, 0.013742079734802246, 0.013648896217346192, 0.013675552368164062, 0.013744095802307128, 0.01440665626525879, 0.01619865608215332, 0.016926719665527345, 0.014181376457214356, 0.01386188793182373, 0.013735936164855958, 0.01376153564453125, 0.013664256095886231, 0.013693951606750488, 0.013689855575561523, 0.013661184310913087, 0.013678591728210449, 0.013649920463562011, 0.014139391899108887, 0.013820927619934082, 0.013752320289611816, 0.013699071884155273, 0.014027775764465332, 0.01368064022064209, 0.013682687759399414, 0.013675519943237305, 0.013695039749145509, 0.01367750358581543, 0.01376460838317871, 0.014016511917114258, 0.0136878080368042, 0.013678591728210449, 0.01369600009918213, 0.013848575592041015, 0.013717503547668456, 0.0136878080368042, 0.013729791641235351, 0.013838335990905762, 0.028457984924316407, 0.013737983703613281, 0.013685759544372558, 0.013691904067993164, 0.013822976112365723, 0.013697024345397948, 0.013604864120483399, 0.013649920463562011, 0.013724672317504882, 0.013774847984313965, 0.013691904067993164, 0.01366528034210205, 0.013736960411071777, 0.013702143669128418, 0.0136878080368042, 0.013701120376586913, 0.013822976112365723, 0.013692928314208984, 0.013674495697021484, 0.013799424171447755, 0.013674495697021484, 0.01368166446685791, 0.013725695610046386, 0.013814784049987794, 0.013684736251831055, 0.01368883228302002, 0.013848608016967773, 0.013685728073120117, 0.013669376373291015, 0.013674495697021484, 0.01367961597442627, 0.01367244815826416, 0.013739007949829102, 0.01367961597442627, 0.013712384223937989, 0.013697024345397948, 0.013697024345397948, 0.013740032196044923, 0.01376460838317871, 0.013657088279724122, 0.013855744361877441, 0.013742079734802246, 0.013683712005615235, 0.013691904067993164, 0.013676544189453126, 0.013682687759399414, 0.013661184310913087, 0.013661184310913087, 0.013647904396057129, 0.013673439979553222, 0.013668352127075196, 0.013643775939941406, 0.013797375679016113, 0.013658111572265624, 0.013664256095886231, 0.013652031898498534, 0.01367033576965332, 0.013643775939941406, 0.013650943756103515, 0.013666303634643554, 0.013641728401184081, 0.01367142391204834, 0.013726719856262207, 0.028403711318969727, 0.013668352127075196, 0.013674495697021484, 0.01363046360015869, 0.013641728401184081, 0.013643775939941406, 0.01365503978729248, 0.01366220760345459, 0.01366220760345459, 0.01368166446685791, 0.013640704154968262, 0.013685759544372558, 0.013637632369995116, 0.013725695610046386, 0.013702143669128418, 0.013674495697021484, 0.013714431762695312, 0.013770751953125, 0.013796352386474609, 0.013739007949829102, 0.013711359977722168, 0.01367244815826416, 0.013684736251831055, 0.013736960411071777, 0.01367244815826416, 0.013675519943237305, 0.01368992042541504, 0.01368569564819336, 0.01368166446685791, 0.01367961597442627, 0.0136878080368042, 0.01367961597442627, 0.013676544189453126, 0.013953023910522461, 0.013818880081176758, 0.01379532814025879, 0.013685759544372558, 0.01367347240447998, 0.013747200012207032, 0.013718527793884277, 0.013677568435668945, 0.013668352127075196, 0.013685759544372558, 0.01365503978729248, 0.013649920463562011, 0.013637632369995116, 0.013633536338806153, 0.013768704414367675, 0.01365401554107666, 0.013760543823242188, 0.013644767761230468, 0.013698047637939453, 0.013644800186157227, 0.013639679908752441, 0.013675519943237305, 0.013670399665832519, 0.013651968002319336, 0.013715456008911133, 0.01376972770690918, 0.013659135818481445, 0.01366528034210205, 0.013699071884155273, 0.013642751693725585, 0.02831974411010742, 0.013700096130371094, 0.013639679908752441, 0.013622271537780761, 0.013687840461730957, 0.013710304260253907, 0.013718527793884277, 0.01368166446685791, 0.01366323184967041, 0.013705216407775878, 0.01368064022064209, 0.013686783790588379, 0.013666303634643554, 0.013668352127075196, 0.013682687759399414, 0.013727744102478028, 0.013724672317504882, 0.013721599578857421, 0.013724672317504882, 0.014102527618408203, 0.01380352020263672, 0.013740032196044923, 0.013737983703613281, 0.013920255661010742, 0.013740032196044923, 0.013717503547668456, 0.013843456268310546, 0.01380352020263672, 0.013684736251831055, 0.013757439613342285, 0.013758463859558106, 0.013974528312683105, 0.013858816146850587, 0.01417420768737793, 0.013920255661010742, 0.013936639785766602, 0.01377280044555664, 0.013703167915344238, 0.01368883228302002, 0.013677568435668945, 0.013684736251831055, 0.013694975852966309, 0.013688863754272461, 0.013667296409606934, 0.013786111831665039, 0.013828096389770507, 0.013668352127075196, 0.013664256095886231, 0.013733887672424316, 0.013702143669128418, 0.013656064033508301, 0.013676544189453126, 0.01377894401550293, 0.013708288192749024, 0.0136878080368042, 0.013683712005615235, 0.013735936164855958, 0.01395404815673828, 0.013817855834960938, 0.013650943756103515, 0.013703167915344238, 0.013696063995361328, 0.013751232147216796, 0.028461055755615236, 0.013700096130371094, 0.013645824432373046, 0.014017536163330077, 0.013718527793884277, 0.015047679901123047, 0.014234623908996581, 0.014027775764465332, 0.01408512020111084, 0.014048255920410157, 0.014173184394836426, 0.013982720375061035, 0.014036992073059081, 0.013888511657714844, 0.013721599578857421, 0.01374617576599121, 0.013685759544372558, 0.013708288192749024, 0.013652992248535157, 0.013644800186157227, 0.013687871932983398, 0.013667263984680177, 0.01366220760345459, 0.01367347240447998, 0.01367961597442627, 0.013639679908752441, 0.01368166446685791, 0.013686783790588379, 0.013645824432373046, 0.013686783790588379, 0.01366220760345459, 0.013650943756103515, 0.013632512092590332, 0.013693951606750488, 0.013690879821777344, 0.013845503807067871, 0.013686783790588379, 0.013660160064697266, 0.013715456008911133, 0.01387724781036377, 0.014060544013977052, 0.01405951976776123, 0.013849599838256836, 0.013645824432373046, 0.013682687759399414, 0.013699071884155273, 0.013699071884155273, 0.013707263946533203, 0.013682687759399414, 0.013713408470153808, 0.013652992248535157, 0.013708288192749024, 0.01368166446685791, 0.01368064022064209, 0.013719552040100098, 0.013873151779174805, 0.01370419216156006, 0.013788224220275879, 0.013749183654785157, 0.013675519943237305, 0.01366329574584961, 0.01380140781402588, 0.014252032279968262, 0.029146112442016602, 0.013622271537780761, 0.013830143928527832, 0.013846528053283692, 0.013924351692199707, 0.013684736251831055, 0.013667327880859375, 0.01367347240447998, 0.01363865566253662, 0.01366323184967041, 0.013697024345397948, 0.013644800186157227, 0.013669376373291015, 0.01386086368560791, 0.013948927879333496, 0.013683712005615235, 0.013719552040100098, 0.013699071884155273, 0.013691904067993164, 0.013702143669128418, 0.013697024345397948, 0.013853728294372558, 0.013726688385009766, 0.013720576286315917, 0.013678591728210449, 0.01369600009918213, 0.013735936164855958, 0.013697024345397948, 0.013715456008911133, 0.013755392074584961, 0.013735936164855958, 0.013721599578857421, 0.013707263946533203, 0.013883392333984374, 0.013845503807067871, 0.013723648071289063, 0.013712384223937989, 0.013705216407775878, 0.01368883228302002, 0.013685759544372558, 0.013721599578857421, 0.013728768348693847, 0.013697024345397948, 0.013751296043395997, 0.013793279647827148, 0.013677568435668945, 0.01369600009918213, 0.013702143669128418, 0.013920255661010742, 0.013789183616638183, 0.013728768348693847, 0.013712384223937989, 0.01368166446685791, 0.013733920097351074, 0.013666272163391113, 0.01387724781036377, 0.013720576286315917, 0.013702143669128418, 0.013740032196044923, 0.013721599578857421, 0.013722623825073242, 0.013712384223937989, 0.01376153564453125, 0.02837708854675293, 0.01365401554107666, 0.01364684772491455, 0.013660160064697266, 0.013664256095886231, 0.013692928314208984, 0.013957119941711426, 0.01405951976776123, 0.013983743667602539, 0.01367961597442627, 0.013649920463562011, 0.013858880043029785, 0.013722559928894044, 0.013641728401184081, 0.013668352127075196, 0.013664256095886231, 0.013723648071289063, 0.01365503978729248, 0.013669376373291015, 0.01367142391204834, 0.01370419216156006, 0.01368166446685791, 0.01367142391204834, 0.013675519943237305, 0.013628416061401367, 0.013627391815185547, 0.013661184310913087, 0.013818880081176758, 0.013710335731506347, 0.013647871971130371, 0.013670399665832519, 0.01368064022064209, 0.01367347240447998, 0.013683712005615235, 0.01368883228302002, 0.013701120376586913, 0.013677568435668945, 0.013657088279724122, 0.013692928314208984, 0.013644800186157227, 0.013676544189453126, 0.013677568435668945, 0.01380352020263672, 0.013749247550964355, 0.013674495697021484, 0.013695039749145509, 0.013658047676086426, 0.013690879821777344, 0.01367244815826416, 0.013657088279724122, 0.013640704154968262, 0.013657088279724122, 0.013793279647827148, 0.013695008277893066, 0.013856736183166504, 0.013647871971130371, 0.013661215782165528, 0.013750240325927735, 0.013722623825073242, 0.013768704414367675, 0.013710335731506347, 0.013717503547668456, 0.01366220760345459, 0.028322816848754883, 0.013648896217346192, 0.013686783790588379, 0.01369600009918213, 0.013699071884155273, 0.0136878080368042, 0.013691904067993164, 0.013702143669128418, 0.013697024345397948, 0.013683712005615235, 0.013700096130371094, 0.013693951606750488, 0.013722623825073242, 0.013705216407775878, 0.013691935539245605, 0.013717472076416016, 0.013724672317504882, 0.013700096130371094, 0.013664256095886231, 0.013639679908752441, 0.013757439613342285, 0.013656064033508301, 0.013667327880859375, 0.01366431999206543, 0.013849535942077636, 0.013924351692199707, 0.013633567810058594, 0.013722592353820801, 0.013668352127075196, 0.013906944274902343, 0.013789183616638183, 0.013662240028381348, 0.013687775611877442, 0.013670399665832519, 0.01367347240447998, 0.01366528034210205, 0.013657088279724122, 0.013661184310913087, 0.013678591728210449, 0.013684736251831055, 0.013658143997192384, 0.013722592353820801, 0.01366528034210205, 0.013637632369995116, 0.01367347240447998, 0.01365503978729248, 0.01367347240447998, 0.013744128227233888, 0.01370419216156006, 0.013693951606750488, 0.013711359977722168, 0.01368064022064209, 0.01368064022064209, 0.013691904067993164, 0.01369600009918213, 0.013724672317504882, 0.013697024345397948, 0.01367142391204834, 0.013682687759399414, 0.01365503978729248, 0.013683712005615235, 0.013806591987609864, 0.013955072402954101]",tokens/s,71.4755369986145,,,,,,,, @@ -32598,7 +32568,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e43c6-0889c0707ee1fb49167cbf42;01cfe96d-6852-4042-9dd2-e86a8618ebe3) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e72e-114c6b7241977a111410b171;39928736-5b45-4f22-b3c4-1d5211c15515) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -32687,7 +32657,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5322-419b4d272fbfb497634be370;3680127b-e14c-4998-b137-f33376772039) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6671007f-04b6d0735fdce28c3de5760e;eea51f2a-ce68-489f-805a-bee3c6bb13a6) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -32747,7 +32717,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e436f-72d668176c6905f840768179;4c00cc37-bf99-4435-b8f2-3db3ed4be6ae) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e6dc-7ba5961c41ce74b570980386;975ebbff-af78-4a18-bcef-d4d8339b9225) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -32979,7 +32949,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp1b8xof9t/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp9b3lpcol/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -33059,7 +33029,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e3e66-4d0314cb6048323116892481;a186d5b0-4510-44dd-9a01-257ec7f28860) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e1a3-7a1c9192775fe3bb190122cb;3b2ff019-adb6-43d7-9fae-18f4d17b887b) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -33144,7 +33114,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5217-4100d6c2791a35860e8c4ff9;4daba840-850e-43cd-bad3-e2351a4781f3) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670ff64-06a28ee53e5ad3c86cf11048;bdb3e053-fd8c-4f6d-a927-a8023a159946) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -33260,7 +33230,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5069-6ff84b9b6becb4a32dfe6d79;e701e34c-2edb-4a06-86d4-8c363df8b618) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fd58-30e61f2617e753c30615bd41;694608b0-707d-495e-912c-f8058cba4318) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -33360,7 +33330,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpw03h_urv/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpjsfiml91/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -33432,7 +33402,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e54c3-0de8569051a3392f2b3e2ddf;029b5649-0142-47d6-9511-f2ad49423f98) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6671021d-6428458615ffd632351434bf;44efecbc-cd51-4c21-b5e9-4bce75f9e578) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -33489,7 +33459,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp2v12tnxy/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp0uegjlf1/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,M,M,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -33531,7 +33501,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5015-20af5c5776dd41167ee09374;72b7d2ee-0092-4f0d-aacc-80d8f812132c) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fcf9-0278cf33557c94430781081a;cf534497-01c4-499b-94d2-ae1a33eb00c3) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -33602,7 +33572,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e516d-1ad4d3a82668d3ef67750ce3;6165642a-2232-4496-8ecc-bf2ef6f414cc) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fe97-2004c1ae40cb31df401e4edb;df96d878-da34-454a-8d46-2ed447016c44) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -33660,7 +33630,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpvj7z6itk/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp_si5nn7c/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,MB,1743.122432,22129.672192,0.0,21483.225088,20799.036928,s,10,27.909502197265624,2.7909502197265623,0.0030225550178342967,2.7914050292968753,2.79439267578125,2.794928149414062,2.7953565283203123,"[2.794273681640625, 2.795463623046875, 2.7873525390625, 2.789448486328125, 2.787631103515625, 2.78628466796875, 2.790373291015625, 2.792436767578125, 2.79304931640625, 2.793188720703125]",tokens/s,91.72503263962948,kWh,3.290233454770512e-05,1.803171549396211e-05,0.0001583880155991868,0.00020932206564085407,tokens/kWh,1222995.765956342,MB,1743.720448,22129.672192,0.0,21483.225088,20902.142976,s,10,1661.799328125,166.1799328125,0.024199823300055656,166.1772734375,166.2117125,166.21745,166.22204,"[166.17496875, 166.167359375, 166.14703125, 166.2104375, 166.2231875, 166.181296875, 166.179578125, 166.204078125, 166.159546875, 166.15184375]",tokens/s,0.37910714569299164,kWh,0.0019616428730719624,0.001075153456676562,0.009382404450362024,0.012419200780110549,tokens/kWh,5072.790199261051,,s,629,1684.2762666015624,2.677704716377683,0.33156351074934576,2.63752294921875,2.6395720703125,2.6405790039062498,5.428127246093751,"[2.63648046875, 2.637477783203125, 2.636610595703125, 2.63773388671875, 2.63674365234375, 2.6374482421875, 2.63684716796875, 2.637740966796875, 2.636241943359375, 2.6376162109375, 2.636966796875, 2.6365615234375, 2.639180908203125, 2.637003662109375, 2.637414306640625, 2.638074951171875, 2.638488525390625, 2.637603759765625, 2.6381865234375, 2.6367314453125, 2.63737646484375, 2.636360595703125, 2.63697314453125, 2.6366044921875, 2.6370693359375, 2.636568603515625, 2.636980224609375, 2.636353515625, 2.63706103515625, 2.636314697265625, 2.637646728515625, 2.63864208984375, 2.637822998046875, 2.63769091796875, 2.636580810546875, 2.636631103515625, 2.6366474609375, 2.636380126953125, 2.635778076171875, 2.639795166015625, 2.637285400390625, 2.636735595703125, 2.63680419921875, 2.636505126953125, 2.636515380859375, 2.637959228515625, 2.6366884765625, 2.638927001953125, 2.6396015625, 2.6393056640625, 2.640069580078125, 2.638898193359375, 2.6396181640625, 2.639129638671875, 2.6404248046875, 2.63950439453125, 2.6398291015625, 2.639203369140625, 2.639447021484375, 2.638824462890625, 2.638972900390625, 2.637701171875, 5.43399951171875, 2.63682861328125, 2.637526123046875, 2.637077392578125, 2.636991455078125, 2.636851318359375, 2.63690966796875, 2.636982177734375, 2.638088134765625, 2.63716455078125, 2.63853662109375, 2.638309326171875, 2.63889111328125, 2.64060107421875, 2.637055908203125, 2.63606884765625, 2.63802880859375, 2.637928466796875, 2.6390927734375, 2.6384384765625, 2.637506591796875, 2.63756298828125, 2.63665869140625, 2.636583984375, 2.63687060546875, 2.6370908203125, 2.637446044921875, 2.63802685546875, 2.6375556640625, 2.637516845703125, 2.63621728515625, 2.637552734375, 2.637285400390625, 2.637663330078125, 2.6371767578125, 2.637763671875, 2.63699560546875, 2.63773583984375, 2.637076416015625, 2.637444091796875, 2.636642333984375, 2.636317626953125, 2.63705078125, 2.636695556640625, 2.637413330078125, 2.63615478515625, 2.6434365234375, 2.637322265625, 2.638057373046875, 2.63786181640625, 2.638676025390625, 2.637177734375, 2.63716455078125, 2.637470703125, 2.63781689453125, 2.636496826171875, 2.63722802734375, 2.637602783203125, 2.638507080078125, 2.637663330078125, 2.636621826171875, 2.63727001953125, 2.6375107421875, 5.42925830078125, 2.638592041015625, 2.6379560546875, 2.637390869140625, 2.636768310546875, 2.63889501953125, 2.638235595703125, 2.638288818359375, 2.637874267578125, 2.638148681640625, 2.637728759765625, 2.637327392578125, 2.63690869140625, 2.637347900390625, 2.6369453125, 2.637433837890625, 2.63707958984375, 2.636822509765625, 2.637656982421875, 2.63617529296875, 2.636454833984375, 2.6358466796875, 2.636423095703125, 2.63856640625, 2.63604833984375, 2.63657470703125, 2.637000732421875, 2.637095947265625, 2.636664794921875, 2.6381435546875, 2.640372802734375, 2.638454833984375, 2.636282958984375, 2.636675048828125, 2.635629638671875, 2.636507080078125, 2.63680615234375, 2.63701611328125, 2.636884033203125, 2.63752294921875, 2.6366669921875, 2.63699462890625, 2.63780859375, 2.6374072265625, 2.636739501953125, 2.6374296875, 2.637625244140625, 2.639678466796875, 2.63699658203125, 2.63729052734375, 2.637656982421875, 2.636547119140625, 2.6364794921875, 2.636135498046875, 2.636613525390625, 2.636599365234375, 2.63747265625, 2.6357861328125, 2.63600439453125, 2.6359716796875, 2.636907470703125, 2.636198974609375, 2.639195068359375, 5.42974560546875, 2.637918212890625, 2.638065673828125, 2.6382080078125, 2.639459228515625, 2.638834716796875, 2.640337890625, 2.639097900390625, 2.638284912109375, 2.639035400390625, 2.636672119140625, 2.63703955078125, 2.63772265625, 2.639283203125, 2.6380830078125, 2.639520751953125, 2.63948291015625, 2.637526123046875, 2.638147705078125, 2.637602783203125, 2.63681640625, 2.63762841796875, 2.638529541015625, 2.63696484375, 2.63714599609375, 2.637178955078125, 2.6376796875, 2.63855810546875, 2.641314697265625, 2.637876220703125, 2.63853369140625, 2.63912548828125, 2.637345703125, 2.636560302734375, 2.636971923828125, 2.636669921875, 2.636619873046875, 2.6392392578125, 2.63902099609375, 2.638465087890625, 2.63775341796875, 2.638636962890625, 2.637673583984375, 2.6372373046875, 2.64167431640625, 2.6387353515625, 2.637962158203125, 2.639416259765625, 2.637681640625, 2.639564697265625, 2.6377236328125, 2.638668701171875, 2.638095458984375, 2.63796728515625, 2.637655029296875, 2.637210693359375, 2.63780859375, 2.638158935546875, 2.637846435546875, 2.63988232421875, 2.641334228515625, 2.636745849609375, 2.63619482421875, 5.42609521484375, 2.63847412109375, 2.64098095703125, 2.6405458984375, 2.64087451171875, 2.64148388671875, 2.64074755859375, 2.64026318359375, 2.640604248046875, 2.640962646484375, 2.638622802734375, 2.63895458984375, 2.636958740234375, 2.636030029296875, 2.636072021484375, 2.637117431640625, 2.636577880859375, 2.642241455078125, 2.64078857421875, 2.641033203125, 2.641383544921875, 2.64137841796875, 2.639701904296875, 2.640848876953125, 2.640034912109375, 2.6397451171875, 2.640384033203125, 2.637106201171875, 2.636432373046875, 2.636267578125, 2.637424560546875, 2.636958740234375, 2.63680712890625, 2.637274169921875, 2.638011474609375, 2.63804931640625, 2.63707861328125, 2.636590087890625, 2.63739794921875, 2.63689013671875, 2.63980029296875, 2.639002685546875, 2.638836669921875, 2.63720849609375, 2.637477783203125, 2.637814697265625, 2.637075439453125, 2.637104248046875, 2.638521240234375, 2.63874755859375, 2.63686962890625, 2.63682763671875, 2.637802490234375, 2.63747998046875, 2.6362255859375, 2.63657568359375, 2.636940185546875, 2.642130859375, 2.6372998046875, 2.63684912109375, 2.63803271484375, 2.63796533203125, 2.637073486328125, 5.4244033203125, 2.636745849609375, 2.6364814453125, 2.6371123046875, 2.637408203125, 2.636239990234375, 2.63697607421875, 2.638215087890625, 2.639066162109375, 2.63918994140625, 2.639048583984375, 2.63975634765625, 2.639510498046875, 2.63817626953125, 2.637822021484375, 2.6385029296875, 2.637655029296875, 2.637551513671875, 2.636876708984375, 2.63750048828125, 2.63714599609375, 2.63798779296875, 2.63727099609375, 2.637442138671875, 2.636712890625, 2.637854736328125, 2.638200927734375, 2.6394736328125, 2.63754541015625, 2.636971923828125, 2.637844482421875, 2.63727197265625, 2.6361865234375, 2.636801025390625, 2.637619140625, 2.637947998046875, 2.63809228515625, 2.637560791015625, 2.6417490234375, 2.63729052734375, 2.63712158203125, 2.63689111328125, 2.63716259765625, 2.635864990234375, 2.6373251953125, 2.63663623046875, 2.63742578125, 2.637765625, 2.63775341796875, 2.638487548828125, 2.638257080078125, 2.63924951171875, 2.639295654296875, 2.639500244140625, 2.63644970703125, 2.637189208984375, 2.637624267578125, 2.639413330078125, 2.639233154296875, 2.63777587890625, 2.637401123046875, 2.637765625, 2.63680712890625, 5.42891748046875, 2.637751220703125, 2.6369423828125, 2.637287353515625, 2.6384208984375, 2.6376162109375, 2.637705078125, 2.6376396484375, 2.636984375, 2.636735595703125, 2.637454345703125, 2.63906005859375, 2.641005615234375, 2.639869873046875, 2.638488525390625, 2.638159912109375, 2.638043212890625, 2.6367958984375, 2.6363330078125, 2.637013916015625, 2.63684716796875, 2.63693115234375, 2.636872802734375, 2.64043212890625, 2.636421142578125, 2.63861962890625, 2.63849169921875, 2.638710693359375, 2.637824951171875, 2.6379111328125, 2.636610595703125, 2.6375966796875, 2.637263916015625, 2.63708984375, 2.637641845703125, 2.637759521484375, 2.6372197265625, 2.63781591796875, 2.63773583984375, 2.638950439453125, 2.637347900390625, 2.63666796875, 2.638234619140625, 2.638180419921875, 2.637719482421875, 2.63872216796875, 2.639690673828125, 2.63703662109375, 2.636780517578125, 2.636748779296875, 2.637224853515625, 2.636233642578125, 2.636378173828125, 2.636863525390625, 2.63923193359375, 2.638593994140625, 2.638180419921875, 2.63895556640625, 2.638899169921875, 2.637263916015625, 2.63600537109375, 2.637727783203125, 2.63825, 5.43155078125, 2.637918212890625, 2.63883056640625, 2.63889208984375, 2.63790185546875, 2.637695068359375, 2.636559326171875, 2.636966796875, 2.63714208984375, 2.639033447265625, 2.6404404296875, 2.640530517578125, 2.64198046875, 2.640649169921875, 2.637720458984375, 2.637758544921875, 2.639655029296875, 2.63736328125, 2.63708251953125, 2.636771240234375, 2.636508056640625, 2.63598583984375, 2.6370068359375, 2.636777587890625, 2.636590087890625, 2.63606884765625, 2.63714111328125, 2.63913671875, 2.637537353515625, 2.63705712890625, 2.636992431640625, 2.638615478515625, 2.636080078125, 2.637097900390625, 2.636854248046875, 2.640509033203125, 2.63925146484375, 2.6389052734375, 2.6383544921875, 2.63872705078125, 2.638085205078125, 2.6382275390625, 2.637551513671875, 2.637571044921875, 2.637200439453125, 2.63806982421875, 2.637642822265625, 2.637718505859375, 2.6373642578125, 2.6366044921875, 2.638297119140625, 2.63918798828125, 2.63943896484375, 2.6392646484375, 2.639478759765625, 2.637856689453125, 2.638085205078125, 2.638904296875, 2.6383564453125, 2.639803466796875, 2.640150634765625, 2.637856689453125, 2.640256103515625, 5.4306796875, 2.637127685546875, 2.6382705078125, 2.640280517578125, 2.639151123046875, 2.639075439453125, 2.63817529296875, 2.638035888671875, 2.638784423828125, 2.639287353515625, 2.637106201171875, 2.637173828125, 2.6368388671875, 2.6376591796875, 2.635720703125, 2.63617333984375, 2.636442626953125, 2.63904150390625, 2.6368818359375, 2.635509765625, 2.636674072265625, 2.6366484375, 2.637701171875, 2.638306396484375, 2.64076806640625, 2.63686865234375, 2.6378896484375, 2.639107177734375, 2.6401259765625, 2.6394111328125, 2.63695166015625, 2.63650927734375, 2.636439453125, 2.63674267578125, 2.637043701171875, 2.637275146484375, 2.637382568359375, 2.63693115234375, 2.636669921875, 2.637486083984375, 2.636729248046875, 2.636916748046875, 2.6372802734375, 2.638066650390625, 2.637075439453125, 2.637048828125, 2.636240966796875, 2.637787109375, 2.636506103515625, 2.636660888671875, 2.63712548828125, 2.637421630859375, 2.637157470703125, 2.636517333984375, 2.637552734375, 2.636777587890625, 2.63666064453125, 2.6360966796875, 2.637487060546875, 2.636655517578125, 2.6372158203125, 2.637021240234375, 2.63826025390625, 5.43210595703125, 2.637992919921875, 2.6383369140625, 2.636921875, 2.637557861328125, 2.637305908203125, 2.63912646484375, 2.63752099609375, 2.6370693359375, 2.638551025390625, 2.6377646484375, 2.638067626953125, 2.637068359375, 2.63798876953125, 2.6374892578125, 2.63644775390625, 2.636535888671875, 2.637727783203125, 2.63798876953125, 2.637382568359375, 2.637177734375, 2.637806640625, 2.63628173828125, 2.6376171875, 2.63735302734375, 2.63743896484375, 2.6397236328125, 2.637740966796875, 2.637970458984375, 2.6363720703125, 2.636199951171875, 2.6363935546875, 2.638331787109375, 2.636755859375, 2.636908447265625, 2.636675048828125, 2.63809033203125, 2.636525634765625, 2.63657373046875, 2.63596435546875, 2.63809130859375, 2.6373671875, 2.636396484375, 2.637010009765625, 2.637177734375, 2.636538818359375, 2.637276123046875, 2.638035888671875, 2.636796875, 2.636739501953125, 2.636539794921875, 2.63828076171875, 2.63714404296875, 2.637462646484375, 2.637360107421875, 2.638277587890625, 2.636430419921875, 2.6370283203125, 2.6365234375, 2.638316650390625, 2.636391357421875, 2.636579833984375, 2.636466064453125]",tokens/s,0.3734541728532224,,,main,False,False,,, @@ -33693,7 +33663,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e5956-049394553f5ac9b04bc4c6c0;746ccd33-b601-4939-b64d-bdd2891296c4) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667106ca-44c54b850331452d463b0e15;2e7bd0e4-bba6-4704-bbdd-bc34f55a43a3) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -33778,7 +33748,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e51c3-7148aefd2844c4376810fd78;5663e127-f803-4220-ba1a-6cf82a153d53) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fef0-3662f1276539ddc4677b51fc;ba193021-466e-41c6-b733-e40c706b2c7b) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -33840,7 +33810,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e431d-2dff06f6794ab8da10bddb67;c0231b4f-f7e4-42f8-b3de-83170f1e01ce) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e687-4fd9ffea42341fcb36ee489a;c14a8081-16eb-4597-b61c-ca057ea3ba10) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -33967,7 +33937,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5119-06ba0e43635c39121edffebd;dd3e12bb-e92d-400d-aa97-5ea96ae38f72) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fe18-0a814e98191646aa38f8b18b;a13f3ea3-617b-4a93-801d-4a0308474845) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -34040,7 +34010,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e441a-42632ca9365ff62b50a4a051;f14ca065-9c47-4904-bf95-18a5c9c4f4c2) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e782-5d749d847fbd384b76f8d681;bbd80371-8636-459d-9b86-afdb64a93e7b) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -34099,7 +34069,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpwa2jvjtu/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpy9xbhinj/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,B,B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -34141,7 +34111,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e541d-1e1d9e46570cef9b23cbc161;c79d62cb-b955-4c29-84c7-a7f4fd58535a) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710177-47b5acaf4e9e7de57e2cf89d;618f0c0c-811f-4028-8e42-932219e569cc) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -34198,7 +34168,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpgvarbknu/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpjjch55hl/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,TencentARC/Mistral_Pro_8B_v0.1,TencentARC/Mistral_Pro_8B_v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,MB,2454.69184,7298.613248,0.0,6652.166144,6323.221504,s,10,7.735312255859376,0.7735312255859375,0.0028429451285762273,0.7729794311523437,0.7762465698242188,0.7783619323730468,0.7800542224121094,"[0.780477294921875, 0.7757764892578125, 0.7711934204101563, 0.770524658203125, 0.7719502563476562, 0.7706873168945313, 0.7725225219726563, 0.7734363403320312, 0.7744036254882812, 0.77434033203125]",tokens/s,330.9497943099635,kWh,9.102796527565035e-06,4.9879214771863185e-06,4.3937086431680494e-05,5.802780443643185e-05,tokens/kWh,4411678.202997362,MB,2454.69184,7298.613248,0.0,6652.166144,6382.564864,s,10,458.5464296875,45.85464296875,0.0073453057278341505,45.852935546875,45.866698828124996,45.8672068359375,45.8676132421875,"[45.8568515625, 45.857828125, 45.85469140625, 45.8511796875, 45.845234375, 45.8491953125, 45.85046484375, 45.84668359375, 45.8665859375, 45.86771484375]",tokens/s,1.3739066738112995,kWh,0.0005414656486152074,0.00029677022395511814,0.002569976393586324,0.003408212266156649,tokens/kWh,18484.764175514054,,s,629,464.77344433593714,0.7389084965595191,0.09178579233429768,0.727773193359375,0.7285446655273438,0.7288094848632812,1.4987512353515624,"[0.72828515625, 0.7276810302734374, 0.7278029174804688, 0.72793701171875, 0.7272877807617187, 0.727383056640625, 0.7272335205078125, 0.7274444580078125, 0.7278551025390625, 0.727720947265625, 0.7276083374023438, 0.7281663818359375, 0.7275867919921875, 0.72804248046875, 0.72800048828125, 0.7276687622070312, 0.7281151733398438, 0.7279267578125, 0.7276728515625, 0.7276615600585937, 0.7278981323242187, 0.7287255249023438, 0.7286405029296875, 0.7275847778320312, 0.7281510620117188, 0.7281069946289063, 0.7279093627929687, 0.7283015747070313, 0.7283804321289062, 0.7276973876953124, 0.7277127685546875, 0.7276400756835938, 0.7282882690429687, 0.7277936401367188, 0.7284705200195313, 0.727784423828125, 0.727952392578125, 0.727930908203125, 0.728369140625, 0.728827880859375, 0.7278551025390625, 0.72924365234375, 0.7275161743164062, 0.7274977416992188, 0.7274014892578125, 0.727947265625, 0.7273707275390625, 0.7273738403320312, 0.7270891723632813, 0.72770458984375, 0.727604248046875, 0.7274721069335938, 0.728226806640625, 0.7281571655273438, 0.7279820556640625, 0.7282749633789063, 0.7275151977539063, 0.7279564208984375, 0.7273912353515625, 0.7277537231445312, 0.7282349853515625, 0.7278714599609375, 1.503824951171875, 0.72740966796875, 0.7277650146484375, 0.7272868041992188, 0.7274219360351563, 0.7278981323242187, 0.7287039794921875, 0.7276185302734375, 0.7274321899414062, 0.7271383056640625, 0.72722021484375, 0.7271024780273437, 0.7274373168945313, 0.7272499389648438, 0.7280332641601562, 0.7276113891601562, 0.7276943359375, 0.7272652587890625, 0.7283251342773438, 0.7285104370117188, 0.7281961059570312, 0.7280137939453125, 0.7283834838867187, 0.7276124267578125, 0.7279022216796875, 0.7284910278320312, 0.7285155639648437, 0.728142822265625, 0.7280394287109375, 0.7274547119140625, 0.7279277954101563, 0.7275509643554687, 0.7277987670898437, 0.7278591918945313, 0.728226806640625, 0.7301795654296875, 0.726887451171875, 0.727372802734375, 0.7274393310546875, 0.7276984252929688, 0.72789404296875, 0.7279503173828125, 0.7277659912109375, 0.7282677612304688, 0.7285678100585937, 0.7279226684570312, 0.7286671142578125, 0.7283712158203125, 0.7282360229492187, 0.727741455078125, 0.7277588500976563, 0.7276482543945313, 0.727857177734375, 0.7283681030273438, 0.7278796997070313, 0.7283937377929688, 0.727920654296875, 0.7278295288085938, 0.728015869140625, 0.7276452026367187, 0.7281694946289062, 0.7286343383789062, 0.7274926147460937, 1.4987724609375, 0.7273533325195313, 0.727457763671875, 0.7278960571289063, 0.7272560424804687, 0.7277005004882813, 0.7277772827148438, 0.7281305541992188, 0.7275653076171875, 0.727499755859375, 0.7276851196289063, 0.7280271606445312, 0.7270768432617187, 0.7270553588867188, 0.7272847290039063, 0.7274874877929688, 0.7273748779296875, 0.727520263671875, 0.727920654296875, 0.7277373657226562, 0.7273338623046876, 0.728322021484375, 0.7278253784179688, 0.7282452392578125, 0.72855859375, 0.7283507080078125, 0.7274547119140625, 0.727931884765625, 0.7276277465820312, 0.7273011474609375, 0.7271905517578126, 0.727736328125, 0.7286302490234375, 0.728036376953125, 0.727710693359375, 0.7286405029296875, 0.728036376953125, 0.7283230590820312, 0.72854833984375, 0.7284182739257813, 0.7284367065429688, 0.729618408203125, 0.7277322387695313, 0.7279042358398438, 0.7280332641601562, 0.7283035888671875, 0.7280281372070313, 0.7276656494140625, 0.7277639770507812, 0.7281858520507812, 0.7287337036132813, 0.7282145385742187, 0.7282606201171875, 0.7274495849609375, 0.727203857421875, 0.7272243041992188, 0.727235595703125, 0.7272662963867188, 0.7274495849609375, 0.727531494140625, 0.7289282836914063, 0.7277404174804688, 0.7281520385742187, 1.498346435546875, 0.7277035522460937, 0.7284224243164062, 0.7275243530273438, 0.727741455078125, 0.727731201171875, 0.7279708251953125, 0.7275581665039063, 0.7280773315429687, 0.7281356811523437, 0.7283568725585937, 0.7276912841796875, 0.7285411987304687, 0.7277701416015625, 0.7284715576171875, 0.72749462890625, 0.7280516967773437, 0.7273963623046875, 0.7273779296875, 0.7273543701171875, 0.7274270629882813, 0.7274116821289063, 0.727689208984375, 0.7282974853515625, 0.7277066040039063, 0.7288258666992188, 0.72807421875, 0.727720947265625, 0.728501220703125, 0.7287817993164063, 0.728057861328125, 0.7277598876953125, 0.72751513671875, 0.7276728515625, 0.7286773681640625, 0.7274321899414062, 0.7279042358398438, 0.727920654296875, 0.727573486328125, 0.7276513061523437, 0.7273072509765625, 0.7275028686523437, 0.7276226806640625, 0.727731201171875, 0.7275745239257813, 0.727709716796875, 0.7274833984375, 0.7272919311523437, 0.7273953247070313, 0.7276328735351563, 0.7272529907226563, 0.7279380493164063, 0.7277659912109375, 0.7278837890625, 0.7274598388671875, 0.7278622436523438, 0.727636962890625, 0.7282718505859375, 0.7279830932617187, 0.7275888671875, 0.72747314453125, 0.7274137573242188, 0.7276810302734374, 1.500190673828125, 0.7274137573242188, 0.727794677734375, 0.72776806640625, 0.7275325317382813, 0.728131591796875, 0.7274035034179688, 0.7273656616210937, 0.727741455078125, 0.7269683227539062, 0.7274475708007813, 0.727498779296875, 0.7274158325195312, 0.7276226806640625, 0.7277803344726562, 0.7272796020507812, 0.7277967529296875, 0.7278120727539062, 0.7281182861328125, 0.7282227172851562, 0.729248779296875, 0.7279073486328125, 0.727394287109375, 0.7274772338867187, 0.7273768920898438, 0.7279462280273438, 0.7280506591796875, 0.7277485961914063, 0.7276800537109375, 0.7273492431640625, 0.7277352905273438, 0.7280148315429688, 0.7277557983398437, 0.7279503173828125, 0.72796875, 0.727677978515625, 0.7273717651367188, 0.7274014892578125, 0.7277393798828125, 0.7273564453125, 0.7273922729492187, 0.7276431274414062, 0.7273267211914063, 0.727183349609375, 0.7276964111328125, 0.7270574340820313, 0.7279892578125, 0.7274506225585937, 0.7283281860351563, 0.7276236572265625, 0.7274864501953126, 0.7277271118164063, 0.7274741821289062, 0.7276431274414062, 0.727984130859375, 0.7277352905273438, 0.727568359375, 0.7290419311523437, 0.7279063110351562, 0.7274669799804687, 0.7279380493164063, 0.7273656616210937, 0.727736328125, 1.4986966552734375, 0.7277783203125, 0.7279697875976563, 0.7273380126953125, 0.7279759521484375, 0.7274772338867187, 0.7276113891601562, 0.72751513671875, 0.727520263671875, 0.7276339111328125, 0.7279185791015625, 0.7278212890625, 0.7281500244140625, 0.7277086791992188, 0.7276032104492187, 0.7277926635742188, 0.7276728515625, 0.7275827026367188, 0.7275140991210938, 0.7280179443359375, 0.7281879272460937, 0.7278212890625, 0.727699462890625, 0.72736767578125, 0.7276728515625, 0.7279083251953125, 0.7273850708007813, 0.7275069580078125, 0.7274024658203125, 0.7273922729492187, 0.7274700927734375, 0.7275847778320312, 0.7282288818359375, 0.7278090209960938, 0.7279667358398437, 0.7283240966796874, 0.7274004516601562, 0.7275899047851563, 0.729038818359375, 0.728173583984375, 0.727530517578125, 0.7272581176757813, 0.727572509765625, 0.7277127685546875, 0.72736767578125, 0.7276503295898438, 0.7283455810546875, 0.72766259765625, 0.7285360717773437, 0.7277281494140625, 0.72745166015625, 0.7276943359375, 0.7278653564453125, 0.7278253784179688, 0.7273615112304688, 0.7273717651367188, 0.7274547119140625, 0.7275899047851563, 0.7274557495117188, 0.7279779663085938, 0.7282565307617187, 0.728158203125, 0.7288606567382813, 1.500291015625, 0.7285186767578125, 0.7282554931640625, 0.727962646484375, 0.72707275390625, 0.7270717163085938, 0.7278837890625, 0.7273738403320312, 0.7277557983398437, 0.7276339111328125, 0.728394775390625, 0.7282606201171875, 0.728369140625, 0.727878662109375, 0.7280169067382812, 0.727794677734375, 0.7282175903320313, 0.7275038452148438, 0.7273502807617187, 0.7271946411132812, 0.7275499267578125, 0.7277998046875, 0.7273502807617187, 0.7283394775390625, 0.7281551513671874, 0.7276728515625, 0.729017333984375, 0.7274383544921875, 0.7284019165039063, 0.7275847778320312, 0.727709716796875, 0.7274024658203125, 0.7274024658203125, 0.7272263793945313, 0.7275448608398437, 0.727203857421875, 0.7272734985351562, 0.7281172485351562, 0.72766259765625, 0.7277854614257813, 0.7278192749023438, 0.7277168579101563, 0.7283128051757812, 0.7274014892578125, 0.7279892578125, 0.7273103637695313, 0.7277905883789062, 0.72785302734375, 0.7278776245117188, 0.727984130859375, 0.7275591430664062, 0.7280732421875, 0.728015869140625, 0.7281817626953125, 0.7277322387695313, 0.72749365234375, 0.7281449584960937, 0.7275447387695313, 0.7287470092773437, 0.727857177734375, 0.727446533203125, 0.7272703857421875, 0.7274813232421875, 1.5015628662109375, 0.7280670776367187, 0.7274475708007813, 0.7279002075195312, 0.7276564331054688, 0.727930908203125, 0.7274649658203125, 0.7278028564453125, 0.7277875366210937, 0.7275980834960938, 0.7279329223632812, 0.7276553955078126, 0.7276932983398438, 0.7277240600585938, 0.7281275024414062, 0.7278305053710937, 0.7283988647460937, 0.727973876953125, 0.7275120849609376, 0.7283138427734375, 0.7280885620117188, 0.728431640625, 0.7279882202148438, 0.7285555419921875, 0.727773193359375, 0.7274137573242188, 0.727572509765625, 0.7277168579101563, 0.7279124755859375, 0.7280169067382812, 0.7274721069335938, 0.7271577758789063, 0.7273502807617187, 0.7277578125, 0.72791552734375, 0.7280302124023438, 0.7272263793945313, 0.727066650390625, 0.7275479125976563, 0.7274383544921875, 0.727099365234375, 0.7272929077148438, 0.7277578125, 0.728215576171875, 0.727783447265625, 0.7276656494140625, 0.72743115234375, 0.7274690551757812, 0.7273277587890625, 0.7281571655273438, 0.7275397338867188, 0.727414794921875, 0.7275867919921875, 0.7271116943359375, 0.727257080078125, 0.727667724609375, 0.7276851196289063, 0.7272703857421875, 0.7279493408203125, 0.7277250366210938, 0.7286610107421875, 0.7273246459960937, 0.7279380493164063, 1.50266162109375, 0.72736767578125, 0.7276441650390625, 0.727141357421875, 0.7274188842773438, 0.7278018798828125, 0.727625732421875, 0.727541748046875, 0.7280783081054687, 0.7277342529296875, 0.72876953125, 0.728369140625, 0.7288411865234375, 0.7279749145507812, 0.7283046264648437, 0.72797900390625, 0.7282718505859375, 0.7281930541992188, 0.728658935546875, 0.7289108276367188, 0.7287183227539062, 0.7283169555664063, 0.7286814575195313, 0.7278428344726563, 0.7285463256835938, 0.72789404296875, 0.72875830078125, 0.72804248046875, 0.7291146240234375, 0.7289763793945313, 0.7286661376953125, 0.728784912109375, 0.728363037109375, 0.7279595336914062, 0.7274700927734375, 0.7275233154296875, 0.7272816772460937, 0.7279974365234375, 0.7284090576171875, 0.72762060546875, 0.728300537109375, 0.727457763671875, 0.7274711303710938, 0.727183349609375, 0.7278960571289063, 0.7272171630859375, 0.7289682006835938, 0.7285933837890625, 0.7276564331054688, 0.7285442504882812, 0.7269273681640624, 0.72795751953125, 0.7275642700195313, 0.7277936401367188, 0.7290480346679687, 0.72764208984375, 0.727804931640625, 0.7280322265625, 0.7278858032226563, 0.727383056640625, 0.7281940307617187, 0.7273421020507812, 0.72789404296875, 1.502066650390625, 0.7280179443359375, 0.7290101928710937, 0.7292303466796874, 0.7280660400390625, 0.728395751953125, 0.72835791015625, 0.72821142578125, 0.7285330200195312, 0.7277250366210938, 0.728616943359375, 0.7284172973632812, 0.7285718994140625, 0.728131591796875, 0.7278837890625, 0.7286558837890625, 0.7279124755859375, 0.7291678466796875, 0.728326171875, 0.7273584594726562, 0.7271588134765625, 0.7275980834960938, 0.7273738403320312, 0.7279912719726562, 0.72772607421875, 0.7297515258789062, 0.7277824096679687, 0.7272632446289062, 0.7278909301757812, 0.7273318481445312, 0.7274424438476562, 0.7272509155273438, 0.7275796508789063, 0.7279483032226562, 0.7278059692382812, 0.7278919677734375, 0.7276881713867187, 0.7284940795898438, 0.7285985107421875, 0.7283128051757812, 0.7282175903320313, 0.7282974853515625, 0.728426513671875, 0.7279165649414062, 0.7284940795898438, 0.7287070922851563, 0.7288955078125, 0.7288780517578125, 0.7277035522460937, 0.72800048828125, 0.72765234375, 0.728056884765625, 0.7274495849609375, 0.7280240478515625, 0.727583740234375, 0.727520263671875, 0.7282698364257812, 0.7273543701171875, 0.7287091064453125, 0.7278960571289063, 0.7277035522460937, 0.7276973876953124, 0.727141357421875]",tokens/s,1.3533475452727455,,,,,,,, @@ -34241,7 +34211,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4d70-5cb9417017003b782e08e9fd;40dbd16d-a662-4cc6-a1e6-60fe3e4e7339) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670f9e3-6ac05628305128406b890b3a;b1770d66-0730-45b6-ac64-59d5e576bdf2) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -34299,7 +34269,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpe0u0vb2_/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpzemqskqe/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemm-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,facebook/opt-350m,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,awq,4,gemm,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,MB,1312.935936,1023.934464,0.0,377.48736,290.348032,s,10,0.6976366119384766,0.06976366119384766,0.0018590846735664541,0.0696058578491211,0.07176354598999024,0.07230553245544434,0.07273912162780762,"[0.07284751892089844, 0.06800761413574219, 0.06769139099121094, 0.06779033660888673, 0.07104889678955079, 0.07164310455322266, 0.07125389099121093, 0.06824649810791016, 0.07096521759033203, 0.06814214324951172]",tokens/s,3669.5321836488733,kWh,8.134236630148721e-07,4.4571909736688417e-07,2.1182893658638067e-06,3.377432126245563e-06,tokens/kWh,75797230.09402885,MB,1312.935936,1023.934464,0.0,377.48736,337.281536,s,10,43.73100244140625,4.373100244140625,0.07142248616314475,4.400765380859375,4.4402288085937505,4.445402490234375,4.449541435546875,"[4.35173583984375, 4.25024462890625, 4.26398681640625, 4.30877587890625, 4.41468994140625, 4.4390791015625, 4.43881396484375, 4.42625927734375, 4.3868408203125, 4.450576171875]",tokens/s,14.40625562709468,kWh,5.082961926562783e-05,2.7857636841584544e-05,0.0001254377571233551,0.00020412501323056748,tokens/kWh,308634.39518232364,,s,629,44.28504171752928,0.07040547172898139,0.008395184351578173,0.07026687622070313,0.07091015625,0.07132590332031251,0.13567237548828126,"[0.0723609619140625, 0.07140863800048829, 0.07176195526123047, 0.07163388824462891, 0.07094169616699218, 0.07094783782958984, 0.07148441314697265, 0.07121202850341797, 0.07107174682617187, 0.07111577606201172, 0.07114342498779297, 0.07076761627197266, 0.07108198547363281, 0.07081574249267578, 0.07077072143554687, 0.07092835235595703, 0.07097344207763671, 0.0714598388671875, 0.07095500946044922, 0.07107481384277343, 0.07127347564697266, 0.07087923431396484, 0.07049420928955077, 0.07076659393310547, 0.07170150756835937, 0.071552001953125, 0.0690360336303711, 0.06716006469726563, 0.06747545623779297, 0.0673433609008789, 0.06707405090332032, 0.06747135925292969, 0.06733312225341796, 0.06719999694824219, 0.06728806304931641, 0.0673095703125, 0.06710784149169922, 0.06706175994873047, 0.0676126708984375, 0.06733926391601562, 0.06688256072998047, 0.06733414459228515, 0.06721126556396484, 0.0674703369140625, 0.06733926391601562, 0.06750822448730469, 0.06778880310058594, 0.06814105224609375, 0.0675225601196289, 0.06693376159667969, 0.06731980895996094, 0.06751334381103516, 0.07044198608398437, 0.0684615707397461, 0.0671098861694336, 0.06701465606689454, 0.06718463897705078, 0.06769561767578125, 0.06716422271728516, 0.06652819061279297, 0.06804582214355469, 0.06770995330810547, 0.13573631286621093, 0.0672204818725586, 0.06833356475830078, 0.06807555389404298, 0.06771196746826172, 0.06805197143554688, 0.06715392303466797, 0.06708428955078125, 0.06746112060546874, 0.06681702423095703, 0.06697062683105469, 0.06744166564941406, 0.06732288360595703, 0.06716006469726563, 0.06725325012207031, 0.06715602874755859, 0.06704019165039063, 0.06741094207763672, 0.06709555053710937, 0.06650572967529297, 0.06745292663574219, 0.06664704132080078, 0.06701261138916016, 0.06722150421142578, 0.06713855743408204, 0.0671098861694336, 0.06744268798828125, 0.06728704071044922, 0.0696258544921875, 0.06788813018798828, 0.06791372680664062, 0.06849024200439453, 0.06736895751953124, 0.06989209747314454, 0.06821990203857421, 0.06771097564697266, 0.06738432312011719, 0.06751948547363282, 0.06726348876953125, 0.06710784149169922, 0.06749183654785157, 0.06747545623779297, 0.06864179229736328, 0.0675563507080078, 0.06683135986328125, 0.06710681915283204, 0.06729011535644532, 0.06674534606933594, 0.06728089904785156, 0.06722354888916016, 0.06724607849121093, 0.06693990325927734, 0.0674703369140625, 0.06723481750488282, 0.06712223815917968, 0.06780818939208984, 0.06794751739501953, 0.06755840301513671, 0.0675758056640625, 0.06719692993164063, 0.0672573471069336, 0.06723481750488282, 0.06736589050292968, 0.13550796508789062, 0.06692147064208985, 0.06734950256347656, 0.07442227172851562, 0.0679557113647461, 0.06727680206298828, 0.06684569549560547, 0.06775091552734375, 0.06815436553955079, 0.06903807830810547, 0.06718975830078125, 0.06693376159667969, 0.06734031677246094, 0.06811030578613281, 0.06972621154785157, 0.06792601776123047, 0.06736486053466798, 0.06915481567382813, 0.06836736297607422, 0.06893875122070313, 0.07033753967285156, 0.06752665710449218, 0.0672194595336914, 0.06725325012207031, 0.06720822143554687, 0.06680572509765625, 0.06687026977539062, 0.0672143325805664, 0.06730445098876953, 0.06715904235839844, 0.06759014129638671, 0.06702899169921875, 0.06747443389892578, 0.06825984191894531, 0.0673280029296875, 0.06705152130126953, 0.06734130859375, 0.06769664001464844, 0.06806425476074218, 0.06706380462646484, 0.06730137634277343, 0.0671825942993164, 0.0676157455444336, 0.06711398315429687, 0.06727577972412109, 0.06933606719970703, 0.06894182586669922, 0.06846259307861328, 0.06729523468017579, 0.06725836944580078, 0.06733312225341796, 0.0672511978149414, 0.0672890853881836, 0.06693376159667969, 0.0676341781616211, 0.06666957092285156, 0.06659993743896485, 0.0665722885131836, 0.0667514877319336, 0.0672368621826172, 0.06741712188720703, 0.06712726593017578, 0.06705049896240234, 0.1353912353515625, 0.0671488037109375, 0.06749900817871093, 0.0674703369140625, 0.06737715148925781, 0.0673064956665039, 0.06735257720947266, 0.06739974212646484, 0.0676075210571289, 0.07006716918945312, 0.06824960327148437, 0.06842777252197266, 0.06737100982666015, 0.06794239807128906, 0.06855276489257812, 0.0674927978515625, 0.06724813079833984, 0.06728294372558594, 0.06814617919921875, 0.06813286590576172, 0.06723583984375, 0.06725222778320313, 0.06753689575195312, 0.06841139221191406, 0.06750617980957031, 0.06819737243652343, 0.06738438415527344, 0.06882195281982421, 0.06757997131347657, 0.06839907073974609, 0.06708633422851562, 0.06791574096679688, 0.06755020904541016, 0.06742527770996094, 0.06748569488525391, 0.066587646484375, 0.06756761932373047, 0.06675154876708984, 0.06733510589599609, 0.06650163269042969, 0.06747647857666016, 0.06739250946044922, 0.06733824157714843, 0.07072870635986328, 0.07021260833740234, 0.07095807647705078, 0.07044915008544922, 0.07039590454101563, 0.07007437133789063, 0.06753177642822265, 0.06737203216552734, 0.07054438018798828, 0.06973747253417968, 0.07040415954589843, 0.07051052856445313, 0.0710830078125, 0.0708136978149414, 0.06919782257080079, 0.06926131439208984, 0.06997299194335938, 0.07050035095214843, 0.0701286392211914, 0.07048499298095703, 0.14131712341308594, 0.06722457885742188, 0.06749183654785157, 0.06760550689697266, 0.07033446502685547, 0.07015731048583984, 0.07011225891113282, 0.06978765106201172, 0.0705269775390625, 0.06983372497558593, 0.06974156951904296, 0.07031603240966797, 0.07043583679199218, 0.0705802230834961, 0.07040614318847656, 0.07013069152832031, 0.0699504623413086, 0.07051058959960937, 0.07024332427978516, 0.06977126312255859, 0.06987673950195313, 0.07046553802490234, 0.07214806365966797, 0.07040809631347657, 0.06932179260253907, 0.07022994995117188, 0.07043071746826172, 0.07018402862548828, 0.07030262756347656, 0.07083622741699219, 0.07088333129882812, 0.06744166564941406, 0.06735155487060547, 0.06759731292724609, 0.06742733001708984, 0.0693780517578125, 0.07071334075927735, 0.07068876647949218, 0.07128985595703125, 0.07055465698242187, 0.07048291015625, 0.07026483154296875, 0.07063346862792969, 0.07024947357177734, 0.0704563217163086, 0.07028530883789062, 0.0700212173461914, 0.07084226989746094, 0.0705269775390625, 0.07041228485107422, 0.07060889434814453, 0.07039590454101563, 0.07050342559814453, 0.0704686050415039, 0.07056486511230468, 0.07071641540527343, 0.07057920074462891, 0.07043276977539062, 0.07038159942626954, 0.07048802947998047, 0.06970470428466796, 0.07084134674072265, 0.07183257293701172, 0.14264422607421876, 0.07059661102294922, 0.0704337921142578, 0.07088127899169921, 0.07030681610107421, 0.07067340850830078, 0.07019110107421875, 0.06985215759277344, 0.07026483154296875, 0.07035391998291016, 0.0703815689086914, 0.07054847717285156, 0.07035903930664063, 0.07019519805908203, 0.0714567642211914, 0.0703662109375, 0.07029964447021485, 0.07025766754150391, 0.07012351989746093, 0.07053107452392578, 0.07036518096923829, 0.07038873291015625, 0.07040204620361327, 0.07045836639404297, 0.07014093017578125, 0.07181926727294922, 0.07061196899414063, 0.07031705474853515, 0.07053209686279296, 0.07042457580566407, 0.07070719909667969, 0.07003033447265625, 0.0699525146484375, 0.07056486511230468, 0.07099187469482422, 0.07034060668945312, 0.07056281280517578, 0.07041228485107422, 0.070940673828125, 0.07032319641113281, 0.0704901123046875, 0.07044915008544922, 0.07032627105712891, 0.07004057312011719, 0.07051776123046875, 0.07030169677734376, 0.07029043579101563, 0.07063859558105469, 0.07095500946044922, 0.07047993469238281, 0.0705125732421875, 0.07045734405517579, 0.07045222473144531, 0.07037542724609375, 0.07044096374511719, 0.0705638427734375, 0.07061196899414063, 0.07037747192382812, 0.07041024017333984, 0.06980403137207031, 0.07040614318847656, 0.07026585388183594, 0.07064780426025391, 0.14216192626953125, 0.07040102386474609, 0.07040819549560547, 0.07026483154296875, 0.07061504364013672, 0.07052496337890625, 0.07054332733154296, 0.07030989074707031, 0.07011020660400391, 0.07049727630615234, 0.07108812713623047, 0.07056179046630859, 0.07050444793701172, 0.07025971221923828, 0.07032217407226563, 0.07038566589355469, 0.07023513793945313, 0.07034368133544922, 0.07049625396728515, 0.07012556457519531, 0.07023721313476562, 0.06932579040527344, 0.07020543670654297, 0.07085977935791016, 0.07015219116210937, 0.07016242980957031, 0.07007129669189453, 0.07072870635986328, 0.07037542724609375, 0.07045836639404297, 0.07033446502685547, 0.07036313629150391, 0.07039180755615235, 0.07009587097167969, 0.07063346862792969, 0.0701839370727539, 0.07108812713623047, 0.07090892791748046, 0.07049727630615234, 0.07038259124755859, 0.07044403076171875, 0.0706519012451172, 0.070761474609375, 0.0694466552734375, 0.0697528305053711, 0.07026380920410157, 0.07064268493652344, 0.0705433578491211, 0.0703477783203125, 0.07133491516113281, 0.0704368667602539, 0.07131238555908204, 0.07170355224609375, 0.0705771484375, 0.07050752258300781, 0.07046451568603515, 0.07079840087890625, 0.07022892761230469, 0.07035084533691406, 0.07033344268798829, 0.0702402572631836, 0.07078604888916015, 0.07056896209716797, 0.13596263122558594, 0.06696959686279297, 0.06721331024169921, 0.06699417877197265, 0.06940467071533203, 0.07078604888916015, 0.07069593811035156, 0.07035187530517578, 0.07053107452392578, 0.07046246337890626, 0.0706344985961914, 0.07024642944335938, 0.07162467193603515, 0.07080652618408204, 0.07045529937744141, 0.07055359649658204, 0.0706519012451172, 0.07040102386474609, 0.06990335845947265, 0.07026080322265625, 0.07056377410888671, 0.07085670471191406, 0.06976921844482421, 0.07021260833740234, 0.06979583740234375, 0.07011634826660157, 0.0700426254272461, 0.07040716552734375, 0.07058329772949219, 0.0699658203125, 0.07026892852783204, 0.0703272933959961, 0.07030989074707031, 0.07069593811035156, 0.07077683258056641, 0.07038259124755859, 0.07024230194091798, 0.07033548736572266, 0.07012454223632812, 0.07299378967285156, 0.07067033386230469, 0.07059865570068359, 0.07034368133544922, 0.07035699462890625, 0.07052902221679687, 0.070687744140625, 0.06977433776855468, 0.07029862213134766, 0.0704901123046875, 0.07047270202636718, 0.07048191833496094, 0.07052799987792968, 0.0706170883178711, 0.07022182464599609, 0.0703272933959961, 0.070255615234375, 0.07052082824707032, 0.07059967803955078, 0.07068160247802735, 0.0705054702758789, 0.07004364776611328, 0.07048089599609375, 0.07136255645751953, 0.14220700073242187, 0.07030985260009766, 0.07026892852783204, 0.0702033920288086, 0.07120588684082031, 0.07035289764404297, 0.07038361358642578, 0.07045222473144531, 0.070181884765625, 0.07042969512939454, 0.07047577667236328, 0.07076150512695313, 0.07077168273925781, 0.07013990020751953, 0.07032422637939453, 0.07050752258300781, 0.07027200317382812, 0.07128268432617188, 0.070614013671875, 0.07047475433349609, 0.07060889434814453, 0.07019827270507813, 0.07051264190673828, 0.07065702056884765, 0.07134515380859376, 0.07106253051757813, 0.07054950714111329, 0.07042253112792969, 0.07038873291015625, 0.0702740478515625, 0.07026687622070313, 0.06861004638671875, 0.0671272964477539, 0.06711507415771484, 0.0671866226196289, 0.06729523468017579, 0.06710169219970703, 0.06714166259765625, 0.06752764892578125, 0.06734130859375, 0.06720921325683593, 0.06747135925292969, 0.06729933166503907, 0.0674150390625, 0.06712422180175781, 0.06744882965087891, 0.06737612915039062, 0.06713958740234376, 0.06726656341552735, 0.0685823974609375, 0.07029043579101563, 0.07070515441894532, 0.0720742416381836, 0.07064166259765625, 0.07065087890625, 0.07042559814453125, 0.07116806030273437, 0.07063648223876953, 0.07048703765869141, 0.07075430297851562, 0.07056896209716797, 0.07044608306884766, 0.07033036804199219, 0.13839053344726562, 0.07068978881835937, 0.07101030731201172, 0.07077279663085938, 0.07043782043457031, 0.070614013671875, 0.07037542724609375, 0.07067443084716797, 0.070181884765625, 0.07043788909912109, 0.07035391998291016, 0.0702003173828125, 0.07099903869628907, 0.07056690979003906, 0.07074406433105469, 0.07023308563232422, 0.07023616027832032, 0.0702208023071289, 0.070150146484375, 0.0707799072265625, 0.07061199951171875, 0.0703927993774414, 0.07016448211669922, 0.07032115173339844, 0.07010406494140625, 0.07089151763916016, 0.07061913299560547, 0.07181926727294922, 0.07085977935791016, 0.07089663696289063, 0.07025459289550781, 0.07071952056884766, 0.07063139343261719, 0.07043276977539062, 0.07035391998291016, 0.07057920074462891, 0.07042867279052735, 0.07073075103759766, 0.07066726684570312, 0.07058124542236328, 0.07049215698242188, 0.07038668823242188, 0.07035903930664063, 0.07047987365722656, 0.07088025665283203, 0.07053414154052734, 0.07050342559814453, 0.07065395355224609, 0.07077375793457032, 0.07083213043212891, 0.070761474609375, 0.07080242919921875, 0.07068364715576173, 0.0717496337890625, 0.07135539245605468, 0.07119155120849609, 0.07112703704833985, 0.07091506958007812, 0.07129190063476562, 0.07068876647949218, 0.07067545318603516, 0.07053619384765625, 0.07090688323974609]",tokens/s,14.20344151445213,,,,,,,, @@ -34463,7 +34433,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4ccc-1aa47e305334cefc4fca3d4d;bde896e5-1e70-4918-9854-f2a2d04cf15c) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670f939-11c5f55c2e628f906c9557f5;a1ce9bbf-1a2e-492b-a011-8ee1df198099) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -34533,7 +34503,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e55ba-4e36e1274b195a860e23fa67;f546e5c5-baf7-4303-ac64-ffe38bc6440b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710327-4bf63ff066f7bbf5173604db;73ab78b5-27ac-4f90-8a6e-7ba3c92ac9c7) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -34603,7 +34573,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5515-3eeedb6177a75aea7cfd224c;a350077a-03dc-43e4-a443-d64cfb90e400) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6671027c-2896489429cb089157f1b208;7698de13-da98-4afa-b7c8-834cbc7eda1c) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -34703,7 +34673,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e52c8-2ec2ee2276a9e472579f98bf;1755a4a9-98dc-472f-a445-4b21e0b2f679) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710023-18921c10632d4d756d1aa290;4c5c3748-6811-4bf7-a5cb-8933f5487d84) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -34802,7 +34772,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e53c3-124a698f5ba8762c73222b69;f8e16770-b3c2-476c-8c47-9c19993ebe96) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6671011d-1e185c674b5c29a73c014812;1be1d29c-a161-4e82-9667-cbe806031bf0) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -35088,7 +35058,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e43be-60554bbf74d3fbf53fd616bd;6d23a60a-9de6-4bfa-a91f-4c6e4d8e9798) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e727-472478d22ed5fdf47175b388;ea42e177-bf5a-4509-b117-263266d1eb77) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -35293,7 +35263,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e531b-59002026780610643b8d0eca;b2213e1c-4853-4994-86cc-8c29f1814666) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710078-32dbfa5444195f84732cd996;0801f857-5e0a-4042-9fd7-9cc6088941fb) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -35353,7 +35323,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e4367-22e49bd676ebe4326d5ebf4c;76a7caba-2c5a-408d-a0aa-1d5f3281382c) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e6d5-697b9000143baa561bb0c63b;5b48302d-ae1c-40e1-8d69-bb762f070639) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -36068,7 +36038,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e3e5d-0a644ca702edd5ed43439275;53e8d466-8168-451b-a57f-204ef5b07189) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e198-6896c7da462fa3bc045c74ff;ae6c7246-d8ff-4a29-8772-e1d1059d50da) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -36153,7 +36123,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5211-4c5769334ea432f72f54da99;bc427e8f-e1ca-4333-8cd2-9cd6bb139a14) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670ff5b-78e2e3b10362670d26ea37e6;02d4feb3-b7ef-4908-9dc3-2a79a46a2e40) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -36399,7 +36369,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5062-40d90b6f2120ce8864ccbcda;86d9ce79-94f5-47fe-838c-623fd1dd99d6) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fd4f-47399d401a8d9e2144d9f64c;78dbbd1a-77d2-4242-9b54-c156ab831375) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -36647,7 +36617,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e54bc-3fa65f516f6b01924a8a2501;84df4fe7-6310-4d7f-a08d-a4687e3e6740) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710217-549514c826bd05ee30fc3fdf;e246e6ba-6e88-4519-b9e4-dac8426b38f3) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -36746,7 +36716,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e500e-1bdbedb304874060171bbe7f;50ed6ea2-437d-4f5e-8c25-e063ec6efe3d) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fcef-1f1395844d7910a50131db9f;b8b162c0-6677-4cd2-a790-993bb38ce0da) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -36845,7 +36815,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5166-73050d5a587cb9c169d91638;527a64e8-a8d3-4c6e-a9be-c962560158c7) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fe90-146aa9774d9dda79230d6e70;00394450-bf77-463b-90d3-163ca1219d63) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -37008,7 +36978,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e594f-1bedd0c46517206a1ad265bf;eec98200-2023-4074-9f43-348469325e61) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667106c3-5a4dd0415a49a2db06847fde;5163248f-aa78-40ae-87e6-20f299905d85) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -37093,7 +37063,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e51bc-236ec0526e78e3936794554a;851b717c-ae32-4103-8d2a-9f346ed2cefe) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fee8-582498d3261d68950859798c;798ae189-7d24-4497-b025-29658bc6aa7f) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -37217,7 +37187,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e4316-2cc8acbe62e0e83d79f3d637;53aa31f8-e9de-495e-a448-98e54b3e2918) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e681-1d563bdb5b6e8d37599c9db0;0105401b-0c58-406e-a270-8d5b79c04fa0) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -37400,7 +37370,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5112-3bd53a580f468e062328196b;ce53ab77-e45e-434c-b9e8-544507320afc) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fe11-247c46ad6266b211792cba79;e5e8be37-dda6-40e3-bac3-03e8e7ee00e0) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -37563,7 +37533,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4413-51f623e95285b51441be5881;257d5595-3161-43c0-8ba2-e29d5093009c) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e77b-7da1d5df509b2d0b2f177afd;d60ba986-6227-4b35-a525-c87f6b0e46cb) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -37726,7 +37696,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5417-04e1cf7564722b2d30148c8a;b48515fc-3c9d-42e9-832f-4a649aac1c94) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710170-237ad8de2ba4279f380410ef;13491835-93a2-4502-8a42-967519ac5192) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -37864,7 +37834,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4d69-3be073fd58e9f98503bf4681;2c6ea656-ec86-4184-9a83-426a9f354657) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670f9dd-39a579fe0c94cfb1052c96a4;52449825-bb52-4c30-ac38-46346ba83b1c) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -38385,7 +38355,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4cc5-44415fb95701430a7d0849d3;99e0fe0d-858d-4125-bbb3-b5f64b5f4105) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670f932-567e92514bcb6889359442e3;ae363c02-1be3-4d37-a17f-566a3a56f4a9) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -38455,7 +38425,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e55b3-26cef7346d0a93cb7e872de2;5e2ca6ee-815a-4777-9410-93d7bfb933dc) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710320-50d5365b14282be76c7cfa01;e9701714-f001-4670-bddc-abdb6c6e94ca) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -38525,7 +38495,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e550e-32a792bc078241d71141f3be;aaf99c3d-0f97-4e5d-a31b-6e1568df1de6) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710273-3dd7951c7f60e3871e4ba2b9;d3ee6264-b45f-46f6-bde2-a35970d2ee67) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -38653,7 +38623,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e52bb-287f647e04bd030730f83e2d;e2975656-fea6-47a7-a551-e74835aaff4f) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710014-5ea7f0e54f8c038c27ed6cba;79ff527a-d24b-4902-b61b-853f544a278a) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -38752,7 +38722,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e53b6-0019c7e15dbd7fb34e033692;7e62abc3-b55c-49c0-9be0-1917417d97bf) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6671010f-4bf1f28b4e86848b4896dbdc;f4135ae4-f4f6-4990-9bbf-97553d3ee8db) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -39078,7 +39048,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e43b0-743dcfa4248c62d20dcef209;4a359de1-b946-47ae-8574-b7c57544383c) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e719-579872b70a83cf5a367b1ae0;b5796c87-d9c2-47c0-b46f-b36007c306ac) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -39283,7 +39253,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e530d-20e799107ee3b8f5422c95ef;d868ac0d-2fc9-4eeb-88f7-19370e2d9692) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6671006a-4177049a400ba8111a5db244;11a5a45b-711e-4a9f-8979-e457aa09d597) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -39343,7 +39313,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e435a-071bd831097610f1371b324a;2c888c72-91a6-418a-bd0b-d80794e61006) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e6c7-122a192e0baab56d29b6e279;0586cc8f-0e5c-4b16-bbfd-253fa526a575) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -40199,7 +40169,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e3e50-0ad5dd030ec138c07d179baa;e2cdbf39-cf0a-4d06-b6ca-daf878360a08) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e189-1ff7d8a64732daa3690d045a;88f31466-4953-4e1f-a5c2-503b152ab8d4) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -40284,7 +40254,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5203-2625e0d5531dddaf0ae535ea;b0f9e2e7-1845-449e-a60b-009d3e5a3e27) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670ff42-411f45794c8e7bd0522647ed;9ed1c39a-8c31-40a5-a001-a365cfc2cb85) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -40602,7 +40572,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5054-3b3f132339975d9f1bd0cbd0;a14e4c75-f908-4708-898f-c334fdf65cbe) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fd40-45ae23e46862ba403d9fc4cb;335ed619-bbbc-40f4-bd16-898bf61bb7ba) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -40890,7 +40860,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e54ae-2670455409c539e76760b56d;b1a94536-6f1d-4541-9aa5-2cb37da00495) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710207-46676a5e164baeab74f2fde5;fcc8037d-6173-463b-8b26-02607f56128f) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -40989,7 +40959,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4fff-190f42d166194444467522ca;5f670cd2-4b1d-44e8-8193-89178e807c5b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fcdf-3b1847247674bfae4e6f285b;8f2c7aab-1a3c-450e-87a2-edc3067d8b09) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -41088,7 +41058,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5158-3cb93daa6b80669562251fde;188e1751-02eb-4295-9f8a-84d13933340e) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fe7c-212a37515ebeec2b00958fcc;706d9bd5-2b35-49bf-a589-96dfacd13a5e) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -41295,7 +41265,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e5941-270dafa527801ca72365efe5;8c4eb422-2515-4c10-b6fb-2d71a5d601fb) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667106b6-699c534a0235faaa0f18d761;e29b4da5-a509-48c3-9f7c-fb0a58b848db) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -41380,7 +41350,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e51ac-1e4c945f0cf5e683653b6be0;085a7a5a-f9b8-4cd7-97d8-2c96fb2bf5f4) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fed9-44a42ea57e91a7ec1a11a596;f79c739d-58a7-42d9-b5ff-2851f78e8ce9) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -41528,7 +41498,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e4308-54b66ea97cb450ed41db4490;0efdfe4b-f7b6-4005-a7d9-451c4d8aebd4) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e673-528c67bc2780aec35425276b;c4b78131-de0f-4936-a07f-828ae6f8dc93) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -41711,7 +41681,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5104-438d5e2801d3fe650eac5b09;b93decbb-e67e-40f1-b43c-00a60d6d7ab9) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fdff-5d559b086f76701f7db16f22;f1341cb8-5978-44b8-9167-add264bb46be) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -41898,7 +41868,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4405-7ed203720593d55a550ccd0a;f9cffb72-b875-4458-93e5-75851627bc62) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e76d-76f0c2db2446936540086e4f;cbe517e7-88e0-418c-902d-c2256231d558) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -42085,7 +42055,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5409-5046fdd71224916f5faa8b95;ede79c15-e449-45e1-9de4-9f1d93ce4a9a) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710162-5cd04a8e58242a7305f74bbd;25f7e589-5b73-4257-83a9-adcb2ee37ef5) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -42243,7 +42213,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4d59-2e5b09b534a9a13449fef09b;3bfde3d2-014a-4f23-9682-eda6c2e65dbc) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670f9cb-2ccfeae831a36f8d703d2528;ad726c6d-f46c-4728-913b-c859fd3c95bf) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -42794,27 +42764,53 @@ AssertionError: AWQ kernels could not be loaded. Please install them from https: ChildProcessError: Traceback (most recent call last): File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target report = worker(*worker_args) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run - backend: Backend = backend_factory(backend_config) - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 79, in __init__ - self.load_model_with_no_weights() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 255, in load_model_with_no_weights - self.load_model_from_pretrained() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 169, in load_model_from_pretrained - self.pretrained_model = self.automodel_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 558, in from_pretrained - return model_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3626, in from_pretrained - model = cls(config, *model_args, **model_kwargs) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/f363ea8a116b3ea829c7a068ca24bc9d3e668083/modeling_internlm2.py"", line 974, in __init__ - self.model = InternLM2Model(config) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/f363ea8a116b3ea829c7a068ca24bc9d3e668083/modeling_internlm2.py"", line 796, in __init__ - self.layers = nn.ModuleList([InternLM2DecoderLayer(config) for _ in range(config.num_hidden_layers)]) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/f363ea8a116b3ea829c7a068ca24bc9d3e668083/modeling_internlm2.py"", line 796, in - self.layers = nn.ModuleList([InternLM2DecoderLayer(config) for _ in range(config.num_hidden_layers)]) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/f363ea8a116b3ea829c7a068ca24bc9d3e668083/modeling_internlm2.py"", line 598, in __init__ - self.attention = INTERNLM2_ATTENTION_CLASSES[config.attn_implementation](config=config) -KeyError: 'sdpa' + File ""/workspace/optimum_benchmark/benchmark/base.py"", line 68, in run + report = scenario.run(backend) + File ""/workspace/optimum_benchmark/scenarios/inference/scenario.py"", line 117, in run + _ = backend.generate(self.inputs, self.config.generate_kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context + return func(*args, **kwargs) + File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 369, in generate + return self.pretrained_model.generate(**inputs, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context + return func(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 1758, in generate + result = self._sample( + File ""/usr/local/lib/python3.10/dist-packages/transformers/generation/utils.py"", line 2397, in _sample + outputs = self( + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl + return forward_call(*args, **kwargs) + File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/d431f9e4014c895663f30e8942d6e3c1adb6b562/modeling_internlm2.py"", line 1202, in forward + outputs = self.model( + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl + return forward_call(*args, **kwargs) + File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/d431f9e4014c895663f30e8942d6e3c1adb6b562/modeling_internlm2.py"", line 1002, in forward + layer_outputs = decoder_layer( + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl + return forward_call(*args, **kwargs) + File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/d431f9e4014c895663f30e8942d6e3c1adb6b562/modeling_internlm2.py"", line 736, in forward + hidden_states, self_attn_weights, present_key_value = self.attention( + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl + return forward_call(*args, **kwargs) + File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/d431f9e4014c895663f30e8942d6e3c1adb6b562/modeling_internlm2.py"", line 623, in forward + qkv_states = self.wqkv(hidden_states) + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1532, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/nn/modules/module.py"", line 1541, in _call_impl + return forward_call(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py"", line 115, in decorate_context + return func(*args, **kwargs) + File ""/usr/local/lib/python3.10/dist-packages/awq/modules/linear/gemv.py"", line 162, in forward + assert AWQ_INSTALLED, ( +AssertionError: AWQ kernels could not be loaded. Please install them from https://github.com/casper-hansen/AutoAWQ_kernels ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-awq-gemv-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,m,m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,awq,4,gemv,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -42856,7 +42852,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4cb7-2d8d195234a5e2c45884e92b;b535f25e-1bc8-4b78-8834-77f5806e9f41) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670f924-1a4cfbe33eb7e8465015b9ea;fcf143ca-7545-4e88-9e02-0be276e16d4a) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -42926,7 +42922,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e55a5-4c1a108d5c796b0971e46e65;625be288-4727-401b-830d-0b6c873b3b7a) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710312-1baa90763954b8226978b8d4;5e0c5f36-3fbd-495b-9842-5bd8d89e4517) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -42996,7 +42992,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5501-58b42e5a75ff83a617d12812;1c51dc79-7f0a-439b-b2c3-8a9a74c65954) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6671025f-3ef1558f4d31ef4c2d3622c4;2fe51f25-6377-4b44-94c6-7120165f7036) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -43124,7 +43120,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e52ad-4f7da1e0018dfcc36d37c36e;88c56421-5c4a-4be5-a9f4-22dd483cdcd2) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710006-2c35168b1c644e6a6619f67f;d50c6454-82f4-47e5-ac39-bb6bec791d40) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -43229,7 +43225,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e53a7-2b9b54572ee1c86533ee3f7f;748c7f0f-a41b-4614-80dc-98d82fe4219c) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710101-5b7ad0907c5943b543b88ca8;46dc23db-0f5f-4b26-af14-7af2b23a2867) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -43533,7 +43529,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e43a1-2dfa23407a79da2800fcada4;686ee5fb-726c-4be6-a449-9fade5fdacac) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e70c-4802c30d175d903f6955e0b6;445df07e-984b-432d-98db-3c55e2f5c879) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -43756,7 +43752,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e52ff-5a636daf6914e56519cc9b44;d209ea02-53cc-4680-a00e-16f15e88ec6b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6671005c-38ad9d0013da78c16c811103;36875439-b9b1-48fb-abc6-1deb666cd352) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -43816,7 +43812,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e434c-387ac8411c0df6e974f23b3b;62dd2e09-1edb-4fb3-a9bf-f02942200a8f) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e6b9-6f379a7d0095fb18794c0907;59e72f1c-cb7d-43a3-aef7-6410d7ece588) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -44573,7 +44569,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e3e42-4dafa7eb0a56cf1f782a93e5;67396c22-e5e7-421c-830c-001bab352f60) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e178-060fea4c293679383d29f23c;45ed8a21-0966-4100-863a-4784efbea3b9) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -44658,7 +44654,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e51f5-382b64db798553f37b625cfe;ca7d185f-691e-435f-8579-c059f53c6345) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670ff2f-48818ea84c78335902d1888e;16aebcc4-69c6-4041-aeaf-1c4b2c097161) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -44910,7 +44906,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5046-7cf075683436251520cd0d63;6fe3a02e-384b-4ca1-afe8-d1aff818e485) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fd30-642dad427b460100629dc8bb;30727d93-1d95-4e4f-a2f7-ca7c84a8869e) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -45158,7 +45154,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e54a1-765ea740045da02f6d27940f;a93e9ecc-7474-4fb2-90c0-c9fb6effa1fe) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667101f9-2a675f15680a8e07502fcc24;4307e687-ef69-45b0-b823-52089e859b4f) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -45263,7 +45259,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4ff1-7f30d2a06a79d3e6432fa065;3d7bbd8b-7c06-4f53-8865-bbd6f06776b4) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fcd1-0abd8b33721aa8a906452bca;b31c506c-c180-44c8-8700-e060062f82cc) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -45368,7 +45364,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e514a-58eec79c4207d62650908688;3cdee806-6618-4c7f-bcf4-7280ac28c867) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fe5d-09e8c86c7d0ea899749f0460;3069ead8-eee6-4bc0-b4e2-558ab357816b) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -45531,7 +45527,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e5933-71f03bd94b3554471168cd5a;e3c19887-f308-49e9-92ca-ce2c0d6b1aed) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667106a8-001175510ea145b57cb3af39;ba448de0-9588-49e5-82c5-dfc427600564) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -45616,7 +45612,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e519e-135ca1c2187bbf553eeff709;ad7db622-88df-4412-9be4-5e20a30b8982) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670feca-7e6842fb2da9e85447cd7d17;87a1dcf4-20fc-4d68-9106-71820d17b349) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -45712,7 +45708,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666e42fa-7bd37b791b645cf20cee1900;83b361e3-2968-42fb-b83a-54df20f85c57) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e665-7126cd6029566b07769508c3;0f57d13f-287e-456d-8dfc-d5ebe07483b8) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -45907,7 +45903,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e50f6-47b058a3429711007499342c;47cd98b1-26da-4dff-91dc-39af8acb5624) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670fdef-6b6652e21e7cec3406185e19;1cd445be-fa78-43a9-9e1b-d07e3fa27005) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -46082,7 +46078,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e43f7-4d41e03366fdf85f4b663503;7cdeea09-e9ed-4ca3-bbd3-abc232fc8162) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e760-06cf40f6397d60f61fa2fdc6;32fa1b54-1cf1-465d-b175-368bdd0a85ba) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -46257,7 +46253,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e53fb-1f41bb341b0b70ec70071869;995e27e1-3169-46c6-9f72-9f3cfdf44b16) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710155-05819a8877348ae14adcede2;6ae45bd4-a88b-4fe4-8a6d-067e556e9b56) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -46395,7 +46391,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4d4b-3deeb1976cbf35b83d055c65;60160b82-e8cd-49a1-89ea-54a9db5da985) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670f9bc-38a0558a45272661351ea315;65060067-8d80-4bd6-9fd1-04160a88d576) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -46931,7 +46927,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e4ca9-03c191be3720c99e439b7dd5;6e550e8c-745e-41a2-a00c-c26c57a2f7c7) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670f917-60d115173c5611f07ee71cec;972c306c-69af-405f-9772-6159c138ca85) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -47001,7 +46997,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e5597-5cb0fb0b43609bd165a084a7;2240748c-30e1-49de-b298-d4b2a3f3c874) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710304-1748c4d52d485cd91da011a2;9c5bbde7-fb3c-4d3b-a25a-818483e5d793) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -47071,7 +47067,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666e54f3-4bcc820b2d7f628d0757223f;df32a064-edd1-439f-bdc6-3526f85866c9) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66710251-4b0be03f222d60d908921ac3;73d11cd4-1ac7-47d1-8c10-3e7bc56570c7) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`.