diff --git "a/perf-df-gptq-1xA10.csv" "b/perf-df-gptq-1xA10.csv" --- "a/perf-df-gptq-1xA10.csv" +++ "b/perf-df-gptq-1xA10.csv" @@ -67,7 +67,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8e8a-33b411a9254ff50d42cd72e3;9b6c19fc-840e-4d05-95dc-08dab3f62fbf) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e73d-6e78a4de38189e871a8fb8d9;33f1f48a-2eb5-47a5-99ae-2034efbbdf2d) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -138,7 +138,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8f0d-37f051395d5ac5ed3cce5ce8;d7bc308f-a50c-40eb-b882-4b9e156e1e6b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e7b6-0dba1752652b817730e4221d;49a3c763-779d-4a5c-aa75-005d5680c464) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -350,7 +350,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666f8808-34d946d17289e19021bce490;a549b1ff-fdc5-44ed-b357-949c38e26b7d) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d80f-45283a9e7d8507c45acf2b6c;c2883c49-e2aa-4600-bdbb-f887d35ae28f) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -439,7 +439,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8eb4-3387cf2e7c7833cc1c4e6f58;5efcc06c-3ad7-4674-b12e-f16c70c1775b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e766-56326525713b95e37f5d2d73;d29bda65-aa76-488a-ae14-b94899cfd606) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -499,7 +499,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666f87d6-7e90a3b76e62fe0828880b93;73267614-dc53-45a0-ba10-f78daede0c4c) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d7e6-2b07a77e5bd821225e3d22b0;30fa1050-f677-4cf8-8f1b-3c1ae4a38d8a) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -828,7 +828,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666f854a-533e050d32ea20634ca06451;f2469c64-245a-419f-b4cb-348d451082c9) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d64c-3e731a8532df908b7cb6ee4a;faaba032-7046-4954-a6eb-565ec430ccc4) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -913,7 +913,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8e31-2c75ffe064250c871859996a;a1c415a2-82ee-4e96-b9e9-a01f39261f46) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e6ed-6111c8325a5e291e3592c6fa;fec846fa-71c7-4c60-8c0c-97daef167832) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1029,7 +1029,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8d3b-7c72212a55a07ab56d675d87;56a27c2a-7f50-4e82-a26b-aa1a3f85a708) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e624-5e5f0f20462627de0eeb3134;79dd094f-8cd6-4693-a0bc-7ab8a901ddde) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1129,7 +1129,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmphjgv66x8/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpqzab82ew/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -1201,7 +1201,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8f92-6d7937c1477571a144716015;d4740210-0291-4f68-aaa2-bf7f464c2c35) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e83f-0f5844cc62326de7034419c5;f2e8d84d-819c-4f29-bb95-bf337fe739bc) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1258,7 +1258,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpv9df1dmh/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp_51ktme_/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,M,M,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -1300,7 +1300,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8d10-4a4a5d445d787fe511c73010;c55a6047-f9f8-478a-a461-2bf67f407404) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e5fc-2c5d453e387404071f4268b4;7a68e059-8daf-43f9-8ae9-570e4070d813) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1371,7 +1371,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8dd7-141637027d3bb2b8695613c2;d37e4415-c9d0-4b6d-94de-a5e27ab51768) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e69c-507e69c1320f3d3b5ee71618;46fb56ae-bad2-47f3-80ff-f00e9e88be40) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1429,7 +1429,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp1ga6fcoj/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp_91tyht6/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,4706.717696,21498.42944,0.0,20851.982336,20236.60032,s,10,15.640249145507811,1.5640249145507812,0.00168203754809238,1.563786376953125,1.564931494140625,1.5666584350585937,1.5680399877929687,"[1.5630980224609374, 1.5644747314453125, 1.5625794677734375, 1.56453466796875, 1.5645272216796875, 1.5645477294921875, 1.5624769287109375, 1.5628692626953125, 1.5683853759765625, 1.5627557373046874]",tokens/s,163.68025702041214,kWh,1.8462728808323543e-05,1.0117648402811028e-05,9.129585081439906e-05,0.00011987622802553362,tokens/kWh,2135535.9958896274,MB,4706.717696,21498.42944,0.0,20851.982336,20339.706368,s,10,879.7075468749999,87.97075468749999,0.002247932535373388,87.97129296874999,87.973384375,87.973496875,87.973586875,"[87.9672265625, 87.968984375, 87.9679609375, 87.9719375, 87.972140625, 87.973609375, 87.973359375, 87.968734375, 87.9729453125, 87.9706484375]",tokens/s,0.7161470902892214,kWh,0.0010385013203819595,0.0005691906291757731,0.005147801034904198,0.0067554929844619305,tokens/kWh,9325.744271351337,,s,629,892.387736694337,1.4187404398956058,0.1859067280964402,1.39627001953125,1.3965248535156252,1.3967189697265625,2.9604817578125,"[1.39608984375, 1.3963448486328125, 1.3963612060546875, 1.3961072998046875, 1.3964923095703126, 1.3963653564453125, 1.396220947265625, 1.3965025634765624, 1.3960928955078125, 1.3962034912109376, 1.3961072998046875, 1.396042724609375, 1.396294677734375, 1.3963448486328125, 1.3963970947265625, 1.3963560791015626, 1.39631201171875, 1.396305908203125, 1.3963643798828125, 1.3958726806640624, 1.3961134033203124, 1.3959710693359375, 1.3963284912109375, 1.396166748046875, 1.3961839599609376, 1.3961236572265625, 1.3963602294921875, 1.3963294677734375, 1.3962445068359375, 1.396177001953125, 1.3961810302734374, 1.3960733642578125, 1.3959208984375, 1.396359130859375, 1.3958072509765624, 1.39637451171875, 1.396179931640625, 1.396137939453125, 1.39983154296875, 1.396505615234375, 1.3963970947265625, 1.3963028564453126, 1.396326416015625, 1.3964442138671875, 1.3961175537109376, 1.396173828125, 1.396177978515625, 1.3965987548828125, 1.3961728515625, 1.3964912109375, 1.3962935791015625, 1.39636328125, 1.39614208984375, 1.3962025146484376, 1.396126708984375, 1.3964083251953125, 1.3963018798828124, 1.396030517578125, 1.3959925537109374, 1.39601513671875, 1.395998779296875, 1.39626904296875, 2.96171923828125, 1.396209716796875, 1.3962506103515624, 1.3962557373046875, 1.3960858154296876, 1.3961513671875, 1.39597412109375, 1.39620654296875, 1.396126708984375, 1.396136962890625, 1.3962086181640625, 1.396041748046875, 1.39642578125, 1.396341796875, 1.3963878173828126, 1.396432861328125, 1.3963028564453126, 1.395948486328125, 1.3960714111328125, 1.3977159423828125, 1.395883056640625, 1.3964984130859375, 1.3963028564453126, 1.3964410400390626, 1.3959066162109375, 1.3964769287109375, 1.396209716796875, 1.3961768798828125, 1.39627001953125, 1.396104248046875, 1.396179931640625, 1.396157470703125, 1.39639404296875, 1.3961943359375, 1.39627001953125, 1.3962998046875, 1.39653125, 1.3978797607421876, 1.39601513671875, 1.39614111328125, 1.3961656494140624, 1.3962823486328124, 1.396073486328125, 1.3962823486328124, 1.3959659423828126, 1.3963233642578126, 1.3965946044921875, 1.3960704345703125, 1.396041748046875, 1.3960120849609374, 1.396105224609375, 1.3962178955078124, 1.39605810546875, 1.396042724609375, 1.3964912109375, 1.3962711181640626, 1.396250732421875, 1.396602783203125, 1.3962762451171875, 1.396516845703125, 1.396326416015625, 1.398640625, 1.3963304443359374, 2.96060302734375, 1.39647900390625, 1.39653125, 1.3961614990234374, 1.3960181884765626, 1.3962342529296874, 1.39601611328125, 1.3958338623046875, 1.39621484375, 1.3961512451171876, 1.3961553955078125, 1.3965394287109374, 1.39652294921875, 1.3962301025390624, 1.396216796875, 1.39635400390625, 1.3961707763671876, 1.396279296875, 1.39607861328125, 1.3962249755859375, 1.395962890625, 1.3963018798828124, 1.396126708984375, 1.3988648681640625, 1.3960888671875, 1.3964031982421874, 1.3962301025390624, 1.3964400634765626, 1.3959556884765625, 1.3960714111328125, 1.3961871337890626, 1.396095947265625, 1.3961861572265626, 1.3961195068359376, 1.3960130615234374, 1.395988525390625, 1.3964349365234374, 1.396515869140625, 1.3959833984375, 1.396220947265625, 1.3962230224609375, 1.396178955078125, 1.396315185546875, 1.3963509521484374, 1.3963929443359375, 1.3960755615234375, 1.3961195068359376, 1.3990062255859375, 1.396305908203125, 1.3959075927734375, 1.3962926025390625, 1.3961728515625, 1.3962230224609375, 1.39626806640625, 1.3965035400390624, 1.3963131103515625, 1.3960294189453124, 1.3960611572265624, 1.396111328125, 1.396221923828125, 1.39627734375, 1.3961727294921875, 1.3960447998046874, 2.961617919921875, 1.3961595458984375, 1.39620458984375, 1.39593212890625, 1.3961533203125, 1.3960263671875, 1.3961451416015624, 1.3962352294921876, 1.3962076416015625, 1.3960560302734375, 1.3962998046875, 1.3964493408203125, 1.39588720703125, 1.396454345703125, 1.3963284912109375, 1.39639501953125, 1.3963642578125, 1.3962127685546875, 1.396188232421875, 1.396221923828125, 1.396116455078125, 1.3964298095703125, 1.3964974365234375, 1.39631201171875, 1.3961666259765626, 1.396533203125, 1.3962506103515624, 1.396453369140625, 1.3960457763671874, 1.396447265625, 1.396515869140625, 1.39631005859375, 1.396305908203125, 1.3964073486328126, 1.396190185546875, 1.3960980224609374, 1.39657421875, 1.396528076171875, 1.3965926513671876, 1.4005042724609376, 1.396791259765625, 1.3964676513671874, 1.3963724365234376, 1.3963826904296874, 1.3961553955078125, 1.396410400390625, 1.39633251953125, 1.3959403076171875, 1.396326416015625, 1.3962608642578125, 1.3963551025390626, 1.3964451904296875, 1.396463623046875, 1.39631103515625, 1.39641552734375, 1.3962301025390624, 1.396401123046875, 1.3960284423828124, 1.3962137451171874, 1.3964359130859374, 1.3964512939453124, 1.3963233642578126, 1.39616357421875, 2.96169482421875, 1.39649951171875, 1.39685986328125, 1.396106201171875, 1.3962987060546874, 1.396274169921875, 1.3959752197265625, 1.3962496337890624, 1.3964339599609374, 1.39608984375, 1.3961011962890626, 1.3962977294921874, 1.3963929443359375, 1.39652197265625, 1.3963203125, 1.3962506103515624, 1.396452392578125, 1.3964031982421874, 1.3964390869140626, 1.396611083984375, 1.3960335693359376, 1.3959803466796874, 1.395914794921875, 1.3961964111328125, 1.3962803955078125, 1.3961553955078125, 1.3965179443359375, 1.39654248046875, 1.3962127685546875, 1.3964461669921875, 1.396337646484375, 1.3962884521484376, 1.3965230712890624, 1.3961533203125, 1.39621484375, 1.3961461181640624, 1.3964666748046874, 1.396431884765625, 1.3962578125, 1.396168701171875, 1.3964727783203126, 1.4002913818359375, 1.396400146484375, 1.3963284912109375, 1.3961728515625, 1.396716552734375, 1.396369384765625, 1.3964410400390626, 1.3962874755859376, 1.3961717529296875, 1.39602734375, 1.396189208984375, 1.39620458984375, 1.3962147216796874, 1.3963602294921875, 1.396046875, 1.39650048828125, 1.3963643798828125, 1.396154296875, 1.396696044921875, 1.396347900390625, 1.3963560791015626, 1.3960509033203126, 2.96171630859375, 1.3963919677734375, 1.3963355712890626, 1.39618505859375, 1.3962926025390625, 1.396115478515625, 1.3960120849609374, 1.396125732421875, 1.396279296875, 1.3963253173828125, 1.3961246337890625, 1.3961092529296875, 1.3965772705078126, 1.396653076171875, 1.3962025146484376, 1.395925048828125, 1.39633154296875, 1.396427734375, 1.3964052734375, 1.3962967529296875, 1.39624755859375, 1.396263916015625, 1.3963182373046874, 1.396494384765625, 1.3964830322265624, 1.3963775634765625, 1.396209716796875, 1.396411376953125, 1.3961912841796875, 1.39610107421875, 1.39620556640625, 1.39635205078125, 1.3962015380859376, 1.396042724609375, 1.3961400146484375, 1.396209716796875, 1.3963079833984375, 1.3961953125, 1.40067431640625, 1.3964073486328126, 1.396336669921875, 1.3963826904296874, 1.396295654296875, 1.39614208984375, 1.3964093017578125, 1.396348876953125, 1.39626806640625, 1.39626806640625, 1.3963868408203124, 1.3961544189453126, 1.3960079345703125, 1.39607861328125, 1.396130859375, 1.3962720947265626, 1.396485107421875, 1.39627001953125, 1.3966878662109374, 1.396177978515625, 1.3960714111328125, 1.39645751953125, 1.3967821044921875, 1.3963929443359375, 1.396221923828125, 2.9624052734375, 1.39640625, 1.396473876953125, 1.396378662109375, 1.3961103515625, 1.3961666259765626, 1.3961380615234376, 1.3961973876953124, 1.3964195556640624, 1.3963243408203125, 1.39621484375, 1.39620556640625, 1.3961348876953126, 1.3965609130859375, 1.3961318359375, 1.396334716796875, 1.396403076171875, 1.3967677001953125, 1.3963458251953125, 1.39630078125, 1.396005859375, 1.396190185546875, 1.396221923828125, 1.3991322021484376, 1.396274169921875, 1.3965567626953126, 1.3959332275390626, 1.39675439453125, 1.3960386962890625, 1.396147216796875, 1.3964830322265624, 1.3962813720703124, 1.3961563720703125, 1.3962669677734374, 1.3960653076171874, 1.3961922607421875, 1.3961605224609375, 1.3963775634765625, 1.396552734375, 1.39624853515625, 1.396236328125, 1.396547607421875, 1.396465576171875, 1.3962352294921876, 1.3962802734375, 1.3960560302734375, 1.3962137451171874, 1.39620556640625, 1.396074462890625, 1.396125732421875, 1.396263916015625, 1.3962332763671874, 1.396306884765625, 1.396263916015625, 1.39645849609375, 1.396369384765625, 1.3965506591796875, 1.396136962890625, 1.395937255859375, 1.3961185302734376, 1.3962281494140625, 1.3962301025390624, 1.400385498046875, 2.960169921875, 1.3960386962890625, 1.3961124267578124, 1.3962587890625, 1.3961328125, 1.3968507080078125, 1.396115478515625, 1.39623828125, 1.3961759033203125, 1.396401123046875, 1.39621484375, 1.3962935791015625, 1.39642578125, 1.3964779052734375, 1.3963294677734375, 1.3960775146484374, 1.397832763671875, 1.396600830078125, 1.39633154296875, 1.3963253173828125, 1.3960980224609374, 1.3963857421875, 1.3962301025390624, 1.39618408203125, 1.396242431640625, 1.3963233642578126, 1.3963560791015626, 1.3964493408203125, 1.39633154296875, 1.396285400390625, 1.3962772216796875, 1.3962454833984375, 1.3963140869140624, 1.396157470703125, 1.39630078125, 1.3960755615234375, 1.3961595458984375, 1.39652392578125, 1.396315185546875, 1.3964451904296875, 1.3963427734375, 1.3962332763671874, 1.39616357421875, 1.3960242919921875, 1.3964410400390626, 1.396295654296875, 1.396368408203125, 1.396148193359375, 1.3961932373046875, 1.396262939453125, 1.3962987060546874, 1.3963458251953125, 1.396189208984375, 1.3962659912109374, 1.3962874755859376, 1.3962987060546874, 1.3963837890625, 1.3963499755859374, 1.3961451416015624, 1.396453369140625, 1.3965670166015625, 1.3963551025390626, 1.3961656494140624, 2.9656298828125, 1.3966131591796875, 1.396486083984375, 1.396496337890625, 1.3961390380859375, 1.3962127685546875, 1.396168701171875, 1.39631005859375, 1.3962650146484374, 1.39608984375, 1.396369384765625, 1.396193359375, 1.3961533203125, 1.396337646484375, 1.396303955078125, 1.3962352294921876, 1.3963612060546875, 1.3983529052734376, 1.39636328125, 1.396357177734375, 1.396263916015625, 1.3963140869140624, 1.3959822998046876, 1.396474853515625, 1.396347900390625, 1.396595703125, 1.396111328125, 1.3961328125, 1.3962701416015626, 1.3961102294921874, 1.3961707763671876, 1.39612158203125, 1.3963203125, 1.3962547607421876, 1.396042724609375, 1.396252685546875, 1.3960447998046874, 1.396010986328125, 1.3960858154296876, 1.3964871826171874, 1.396463623046875, 1.396033447265625, 1.3962823486328124, 1.3962874755859376, 1.3961461181640624, 1.396464599609375, 1.3964410400390626, 1.396326416015625, 1.396304931640625, 1.396130859375, 1.3960826416015626, 1.396231201171875, 1.39595166015625, 1.3962086181640625, 1.39643505859375, 1.3962188720703126, 1.39656396484375, 1.396241455078125, 1.396109375, 1.396360107421875, 1.3962454833984375, 1.396515869140625, 1.4013460693359374, 2.9601669921875, 1.396242431640625, 1.3961563720703125, 1.3967093505859376, 1.3961451416015624, 1.3960601806640625, 1.3962076416015625, 1.396126708984375, 1.396243408203125, 1.3964246826171876, 1.3964666748046874, 1.39620654296875, 1.39607958984375, 1.39641552734375, 1.396093994140625, 1.3960858154296876, 1.3963212890625, 1.396093994140625, 1.3965445556640625, 1.39627001953125, 1.3963775634765625, 1.396485107421875, 1.3963131103515625, 1.3964759521484376, 1.3963275146484375, 1.3965045166015626, 1.39883935546875, 1.3962230224609375, 1.3962762451171875, 1.396242431640625, 1.3961400146484375, 1.3962178955078124, 1.3964820556640625, 1.396148193359375, 1.3961688232421876, 1.3961829833984376, 1.396322265625, 1.396190185546875, 1.396115478515625, 1.39614208984375, 1.3964031982421874, 1.39636328125, 1.3966151123046875, 1.3964134521484375, 1.3963355712890626, 1.396177978515625, 1.396125732421875, 1.396169677734375, 1.396275146484375, 1.3965240478515626, 1.3966817626953125, 1.3962578125, 1.396337646484375, 1.3967205810546874, 1.3967247314453124, 1.396284423828125, 1.396473876953125, 1.3963509521484374, 1.3962127685546875, 1.3961492919921874, 1.3961031494140625, 1.3963387451171876, 1.39633251953125]",tokens/s,0.7048505645427167,, @@ -1462,7 +1462,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666f9153-2acdd6af4d7b1ccb3e2218de;59e76e18-9bb0-40ac-a469-55a9f325e08e) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e9e3-12b9040642b4ce1a06e877f7;b70095d1-5ed2-4513-87dc-e4896b71b955) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -1547,7 +1547,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8e06-47adcffc59443588053fa103;dabf3f9a-a8d0-4bdb-aaf6-84403ee5e798) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e6c4-76631f0e6ce673311bbbf2fa;0d057160-c793-4984-becc-f079016e6c5c) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1609,7 +1609,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666f87ad-0e600bf30b12b420196cf3c2;4c3a4f89-0d99-43bc-a531-006c49c4ac71) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d7be-7f50158c75e32bc36418b187;eb0c410e-0157-4fda-935a-bed30d952bb1) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -1736,7 +1736,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8dac-1adf52f61f234ab1035fbd32;e0d72a9b-506b-43a5-b816-819cba6c83f9) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e674-6605acc65b6bdd7e098f8bad;90df0b50-2962-4398-af50-399139f8c444) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1809,7 +1809,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8833-261f08cc775ac59903f608c3;6a3c0d23-00eb-4580-80ed-cec923c384f3) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670d837-571681051a23960e1e9aefa3;b7def1cb-7d93-455a-a5d3-6807c4f3fd7a) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1868,7 +1868,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpwg7gv46l/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp53bqdw_e/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,B,B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -1910,7 +1910,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8f3b-51784b96527a7fd344d1ea43;8141f0e8-3f2a-40f1-83b8-6c05e2963a46) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e7df-20a3ef7e5b1e2ba96016b916;f1079b5b-4f5e-42cf-be3b-f376e9c2739f) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1967,7 +1967,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpiwc2zjey/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp1ascikep/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,TencentARC/Mistral_Pro_8B_v0.1,TencentARC/Mistral_Pro_8B_v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,5946.503168,7053.246464,0.0,6406.79936,6098.60864,s,10,4.370441192626953,0.4370441192626952,0.0009281353036040856,0.43696878051757815,0.4381058013916016,0.4383167251586914,0.4384854641723633,"[0.4385276489257813, 0.4359717712402344, 0.43771954345703123, 0.43744598388671874, 0.4380589294433594, 0.4362575378417969, 0.43786093139648435, 0.4364915771484375, 0.436230712890625, 0.43587655639648437]",tokens/s,585.7532196792366,kWh,5.148596204997261e-06,2.8211948946868096e-06,2.4996191494520076e-05,3.296598259420415e-05,tokens/kWh,7765580.754902423,MB,5946.503168,7055.343616,0.0,6408.896512,6157.952,s,10,244.94424023437497,24.494424023437496,0.005272769015065994,24.493744140625,24.502037109375,24.502107421875,24.502163671875,"[24.487064453125, 24.502177734375, 24.49926953125, 24.4968515625, 24.502021484375, 24.493955078125, 24.493533203125, 24.4922109375, 24.487486328125, 24.489669921875]",tokens/s,2.5720139383444343,kWh,0.00028916327015789236,0.00015848618370445323,0.00135038218658508,0.0017980316404474258,tokens/kWh,35038.315557296286,,s,629,248.48397949218753,0.3950460723246224,0.05195367916433446,0.388748291015625,0.38915277099609374,0.38942371215820315,0.8254186791992187,"[0.38863052368164064, 0.3885834350585938, 0.38854656982421876, 0.38854452514648435, 0.3885762634277344, 0.3887677307128906, 0.3885189208984375, 0.3887728576660156, 0.3885434875488281, 0.38884658813476564, 0.38842572021484373, 0.3884328918457031, 0.3885711364746094, 0.3883417663574219, 0.3885240173339844, 0.38941387939453126, 0.3884031982421875, 0.3887196044921875, 0.38854452514648435, 0.3887032470703125, 0.3886264343261719, 0.38861721801757815, 0.3888977966308594, 0.3890206604003906, 0.388917236328125, 0.3884503173828125, 0.388843505859375, 0.3885404052734375, 0.38881793212890625, 0.38851788330078124, 0.3882977294921875, 0.3884666748046875, 0.388701171875, 0.38850970458984374, 0.3885455322265625, 0.3883397216796875, 0.3884103698730469, 0.388495361328125, 0.3894988708496094, 0.3882915954589844, 0.3888158569335938, 0.3886602783203125, 0.3886519775390625, 0.3887790222167969, 0.3884134521484375, 0.38843392944335936, 0.3885281372070313, 0.3888619384765625, 0.38980096435546874, 0.38914764404296875, 0.3887196044921875, 0.38860186767578125, 0.38848306274414063, 0.38838375854492185, 0.3885035400390625, 0.3884605407714844, 0.38845645141601565, 0.38830081176757814, 0.3884513244628906, 0.3889899597167969, 0.3887943725585937, 0.3889407958984375, 0.8262952880859376, 0.38884658813476564, 0.38921319580078123, 0.388748291015625, 0.3887923278808594, 0.38922137451171873, 0.3888977966308594, 0.3885906066894531, 0.3892326354980469, 0.38878106689453124, 0.38923776245117186, 0.38898483276367185, 0.3888721923828125, 0.38877694702148435, 0.38888754272460935, 0.3888670654296875, 0.38904830932617185, 0.3887943725585937, 0.38909030151367185, 0.38900018310546874, 0.3890636901855469, 0.3890831298828125, 0.3892695007324219, 0.38899404907226565, 0.39072256469726563, 0.3887984619140625, 0.3887677307128906, 0.3885977478027344, 0.38868377685546873, 0.38868069458007815, 0.38871551513671876, 0.3887615966796875, 0.38870834350585937, 0.38858139038085937, 0.388885498046875, 0.3888230285644531, 0.3896135559082031, 0.39009588623046876, 0.3886540832519531, 0.38893875122070315, 0.3886039123535156, 0.3890165710449219, 0.3886899108886719, 0.38870529174804686, 0.38863665771484374, 0.3886868591308594, 0.38889166259765623, 0.3890595703125, 0.38893771362304685, 0.3889479675292969, 0.388917236328125, 0.38900222778320315, 0.38909951782226565, 0.38838067626953127, 0.388495361328125, 0.3884666748046875, 0.389074951171875, 0.38862539672851565, 0.38855474853515626, 0.3885578308105469, 0.388843505859375, 0.389080078125, 0.3892039794921875, 0.8267151489257812, 0.3888455810546875, 0.389074951171875, 0.38913229370117186, 0.38876263427734375, 0.3887595520019531, 0.3888875732421875, 0.38918243408203124, 0.3892049865722656, 0.38914047241210936, 0.3887319030761719, 0.38899713134765623, 0.38865509033203127, 0.38901556396484377, 0.3886766052246094, 0.38860800170898435, 0.388864013671875, 0.38877593994140625, 0.3888302001953125, 0.388790283203125, 0.3885578308105469, 0.38857318115234374, 0.3885926513671875, 0.3890616455078125, 0.3892623291015625, 0.3890933837890625, 0.38865817260742186, 0.3889070129394531, 0.38874929809570313, 0.38892340087890626, 0.3887185974121094, 0.3888066711425781, 0.38876568603515627, 0.3885875244140625, 0.38916607666015623, 0.38873394775390624, 0.3885383605957031, 0.38869403076171877, 0.38857217407226563, 0.38951425170898435, 0.38901556396484377, 0.38918759155273436, 0.3889551391601562, 0.3887442016601563, 0.3890032653808594, 0.38889675903320314, 0.3889725341796875, 0.3889837951660156, 0.38876568603515627, 0.38967501831054685, 0.3893606262207031, 0.38946200561523436, 0.3886407775878906, 0.38881381225585937, 0.38867251586914064, 0.38852505493164063, 0.38857217407226563, 0.38876568603515627, 0.38846771240234373, 0.38870529174804686, 0.3890585632324219, 0.38904730224609374, 0.38865716552734375, 0.8268502807617187, 0.3887595520019531, 0.3887820739746094, 0.38875750732421877, 0.3886417846679687, 0.38880767822265627, 0.38890087890625, 0.38849432373046877, 0.38883224487304685, 0.3887718505859375, 0.38920294189453125, 0.3890390930175781, 0.3896995849609375, 0.38909747314453125, 0.3887442016601563, 0.388569091796875, 0.3890595703125, 0.3884718017578125, 0.38869195556640623, 0.38880563354492187, 0.38886605834960936, 0.38868377685546873, 0.3896105041503906, 0.38871450805664065, 0.3889029235839844, 0.38878311157226564, 0.3886878662109375, 0.38854757690429687, 0.38863873291015627, 0.3887370300292969, 0.38861004638671875, 0.38876263427734375, 0.3888015441894531, 0.3884984436035156, 0.38875546264648436, 0.38851174926757814, 0.388959228515625, 0.38882406616210935, 0.3891138610839844, 0.38870733642578126, 0.38867251586914064, 0.38911282348632814, 0.38889166259765623, 0.3884820556640625, 0.38871450805664065, 0.3886540832519531, 0.388600830078125, 0.389032958984375, 0.389180419921875, 0.38953677368164064, 0.38902886962890626, 0.3890667419433594, 0.389001220703125, 0.3887820739746094, 0.38870016479492187, 0.3886346130371094, 0.3887544250488281, 0.38863873291015627, 0.38883328247070315, 0.3885629577636719, 0.3887093811035156, 0.38889675903320314, 0.38913946533203125, 0.8280023193359375, 0.38870834350585937, 0.38864794921875, 0.3887124328613281, 0.388748291015625, 0.3888793640136719, 0.3885383605957031, 0.38862744140625, 0.38867352294921875, 0.3893258361816406, 0.3888793640136719, 0.38871551513671876, 0.38890802001953123, 0.38877490234375, 0.38909747314453125, 0.3890729064941406, 0.3887790222167969, 0.38876876831054685, 0.3889479675292969, 0.38905447387695313, 0.388927490234375, 0.38944769287109376, 0.3890114440917969, 0.3895459899902344, 0.3894302673339844, 0.3890831298828125, 0.38988287353515627, 0.388738037109375, 0.3887667236328125, 0.38885888671875, 0.3889950866699219, 0.3887370300292969, 0.3887564697265625, 0.38875546264648436, 0.3886991271972656, 0.389185546875, 0.38876058959960935, 0.38907086181640627, 0.38890188598632813, 0.3888619384765625, 0.38933401489257813, 0.388853759765625, 0.38896026611328127, 0.38860595703125, 0.38862744140625, 0.38865817260742186, 0.388463623046875, 0.3885875244140625, 0.3883212890625, 0.3891998596191406, 0.3895500793457031, 0.3895326843261719, 0.389032958984375, 0.3888066711425781, 0.388822021484375, 0.388811767578125, 0.38866329956054685, 0.38884146118164065, 0.38888958740234375, 0.38867147827148435, 0.38877490234375, 0.3891199951171875, 0.38873599243164064, 0.8254505004882813, 0.38886605834960936, 0.3887943725585937, 0.3888025512695312, 0.38903604125976565, 0.38877694702148435, 0.38867864990234374, 0.3889029235839844, 0.3888517150878906, 0.38892645263671877, 0.38902374267578127, 0.38975079345703123, 0.389823486328125, 0.389465087890625, 0.3888455810546875, 0.3890862121582031, 0.38899200439453124, 0.388611083984375, 0.3885342712402344, 0.3887175598144531, 0.38892340087890626, 0.38893875122070315, 0.3891138610839844, 0.38856704711914064, 0.3884707946777344, 0.3899740295410156, 0.388389892578125, 0.388780029296875, 0.3886213073730469, 0.3885793151855469, 0.3890093994140625, 0.38884454345703123, 0.3886991271972656, 0.388748291015625, 0.3889151916503906, 0.38854757690429687, 0.3891988525390625, 0.3890667419433594, 0.388822021484375, 0.38893670654296875, 0.3886960754394531, 0.38903399658203125, 0.38867251586914064, 0.38844415283203126, 0.3887595520019531, 0.3886796875, 0.3885424499511719, 0.3885875244140625, 0.3885660095214844, 0.388421630859375, 0.38844927978515625, 0.3885557861328125, 0.3888015441894531, 0.3884236755371094, 0.38832846069335936, 0.388347900390625, 0.388389892578125, 0.38839910888671875, 0.38838580322265626, 0.38835302734375, 0.3885363159179688, 0.38859982299804685, 0.38915277099609374, 0.8270069580078125, 0.388790283203125, 0.38852301025390623, 0.38860800170898435, 0.38870529174804686, 0.3886397399902344, 0.3888097229003906, 0.38888140869140625, 0.38900531005859373, 0.3891435546875, 0.38884454345703123, 0.38879129028320314, 0.38904934692382814, 0.38878924560546874, 0.38913946533203125, 0.38866329956054685, 0.3886970825195312, 0.3890196533203125, 0.38881076049804686, 0.38890087890625, 0.3889612731933594, 0.38866021728515626, 0.3888097229003906, 0.38903604125976565, 0.38920907592773435, 0.389138427734375, 0.38885888671875, 0.38871142578125, 0.38858648681640623, 0.3885793151855469, 0.38836428833007813, 0.38872164916992186, 0.3887953796386719, 0.3888721923828125, 0.389064697265625, 0.3893247985839844, 0.3885936584472656, 0.38854757690429687, 0.3885936584472656, 0.3884789733886719, 0.38885479736328127, 0.38876058959960935, 0.38854656982421876, 0.3885404052734375, 0.38906982421875, 0.38860287475585936, 0.38849945068359376, 0.38863665771484374, 0.38878106689453124, 0.3889336242675781, 0.38914559936523435, 0.38887322998046875, 0.3893893127441406, 0.3887462463378906, 0.38847796630859377, 0.38849127197265626, 0.38849740600585936, 0.3886684265136719, 0.3886510009765625, 0.38863565063476563, 0.38843597412109376, 0.3889039306640625, 0.38858853149414063, 0.8253368530273437, 0.38870220947265627, 0.38867864990234374, 0.3885823974609375, 0.3886929931640625, 0.3884881896972656, 0.38854656982421876, 0.38842059326171874, 0.38854757690429687, 0.38858648681640623, 0.388569091796875, 0.3887452087402344, 0.3891435546875, 0.38859982299804685, 0.3888302001953125, 0.38854144287109377, 0.3885936584472656, 0.38863052368164064, 0.38876980590820315, 0.3885404052734375, 0.38905447387695313, 0.38915277099609374, 0.38901556396484377, 0.38856704711914064, 0.3885404052734375, 0.38846157836914064, 0.38859469604492186, 0.38851071166992185, 0.38895001220703124, 0.38856805419921875, 0.38871450805664065, 0.38837246704101563, 0.3885987854003906, 0.38849127197265626, 0.38861618041992185, 0.38849639892578125, 0.38856192016601565, 0.38879641723632813, 0.38922341918945313, 0.38905752563476564, 0.38980096435546874, 0.3889090576171875, 0.3889070129394531, 0.3885916137695313, 0.38894183349609374, 0.3886796875, 0.3886407775878906, 0.3885383605957031, 0.38921829223632814, 0.38887322998046875, 0.38906265258789063, 0.3887585144042969, 0.38897869873046875, 0.38905548095703124, 0.38911077880859374, 0.3886458740234375, 0.3885455322265625, 0.3886878662109375, 0.3890585632324219, 0.38870220947265627, 0.38884658813476564, 0.38864486694335937, 0.3889561462402344, 0.82583349609375, 0.3893094482421875, 0.38874725341796873, 0.38864697265625, 0.3890738525390625, 0.3888230285644531, 0.3885772705078125, 0.3886264343261719, 0.38845541381835935, 0.38850457763671875, 0.38880767822265627, 0.3886039123535156, 0.3885486145019531, 0.388485107421875, 0.3883673706054688, 0.3887646789550781, 0.3885353088378906, 0.388632568359375, 0.38862539672851565, 0.3885936584472656, 0.38859674072265626, 0.3886069641113281, 0.388485107421875, 0.38892440795898436, 0.38891622924804686, 0.38909133911132815, 0.388453369140625, 0.38876058959960935, 0.3886592102050781, 0.3885567932128906, 0.3885189208984375, 0.3885557861328125, 0.388716552734375, 0.3882700805664063, 0.38850457763671875, 0.3887790222167969, 0.38870733642578126, 0.3885281372070313, 0.38864382934570313, 0.3885977478027344, 0.38863565063476563, 0.3885660095214844, 0.38846463012695315, 0.38862234497070314, 0.3884236755371094, 0.3884984436035156, 0.388347900390625, 0.38830181884765624, 0.38828134155273436, 0.3885066223144531, 0.38897152709960936, 0.38908721923828127, 0.3887431640625, 0.3887175598144531, 0.3884800109863281, 0.38863665771484374, 0.39151205444335935, 0.3889510498046875, 0.38859674072265626, 0.38852301025390623, 0.38847796630859377, 0.3888824462890625, 0.38877593994140625, 0.824754150390625, 0.38834994506835935, 0.38857217407226563, 0.38853323364257814, 0.3888844909667969, 0.3888005065917969, 0.3891640319824219, 0.3885987854003906, 0.3888486328125, 0.38872781372070314, 0.38896435546875, 0.3892623291015625, 0.3890882568359375, 0.3891937255859375, 0.38884658813476564, 0.3889827880859375, 0.38853323364257814, 0.3885260925292969, 0.38840933227539065, 0.38836428833007813, 0.388558837890625, 0.38827621459960937, 0.3885209655761719, 0.3886376953125, 0.38875341796875, 0.3887790222167969, 0.38885888671875, 0.3887298583984375, 0.38882406616210935, 0.38858035278320313, 0.388927490234375, 0.3888988037109375, 0.38889166259765623, 0.38841650390625, 0.3884298095703125, 0.3886039123535156, 0.3884247131347656, 0.3889837951660156, 0.38898175048828126, 0.3884728393554687, 0.3891671142578125, 0.3885987854003906, 0.3886120910644531, 0.388411376953125, 0.38858139038085937, 0.3884390258789063, 0.3885076599121094, 0.388558837890625, 0.388738037109375, 0.38887322998046875, 0.38831716918945314, 0.3885014953613281, 0.3886868591308594, 0.38870220947265627, 0.38861721801757815, 0.38846258544921874, 0.3886673889160156, 0.3884789733886719, 0.38866943359375, 0.388358154296875, 0.3887667236328125, 0.38856500244140624, 0.388558837890625]",tokens/s,2.5313503159658475,, @@ -2010,7 +2010,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8b97-6f5b28f23b17bb203a5d0d61;f3d9ca3b-3f3d-467c-b72c-99f06d4fa7bd) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e48d-429c38ef7d866605540f2936;c0438915-b7b1-4ac9-b8d5-43ed841fa18c) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2068,7 +2068,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpwwaawev7/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmptox0y53l/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,facebook/opt-350m,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1279.791104,1019.74016,0.0,373.293056,277.410816,s,10,0.3806483840942383,0.03806483840942383,0.0010374051807092283,0.03839825630187988,0.038631001281738284,0.03909522857666015,0.03946661041259766,"[0.03955945587158203, 0.03852783966064453, 0.03837411117553711, 0.038422401428222654, 0.03847126388549805, 0.038292991638183595, 0.03834467315673828, 0.03843852615356445, 0.036084991455078125, 0.03613212966918945]",tokens/s,6725.366787229584,kWh,4.3617786259962035e-07,2.3900503699099743e-07,1.1732389459702955e-06,1.8484218455609134e-06,tokens/kWh,138496523.73174343,MB,1280.086016,1019.74016,0.0,373.293056,323.047424,s,10,23.712255126953128,2.371225512695313,0.003882883813537052,2.3718930664062503,2.376249291992188,2.3762570922851562,2.3762633325195313,"[2.362728515625, 2.37624755859375, 2.373291748046875, 2.367230224609375, 2.371970947265625, 2.371815185546875, 2.376264892578125, 2.37161865234375, 2.37223681640625, 2.3688505859375]",tokens/s,26.56854004931377,kWh,2.7777078536237154e-05,1.5222735153430559e-05,6.781520513363374e-05,0.00011081501882330147,tokens/kWh,568514.9961527847,,s,629,24.01549103546141,0.03818043089898478,0.004509392475924156,0.03761459350585938,0.03786444854736328,0.03806208114624023,0.07505829895019543,"[0.03638784027099609, 0.03659775924682617, 0.036880382537841795, 0.03663974380493164, 0.03670937728881836, 0.03717529678344727, 0.036736000061035154, 0.03655987167358398, 0.03629260635375976, 0.03701657485961914, 0.03663974380493164, 0.0363612174987793, 0.03665100860595703, 0.038844417572021485, 0.03839078521728516, 0.038013950347900394, 0.03768627166748047, 0.03772825622558594, 0.037766143798828124, 0.03749990463256836, 0.037585918426513674, 0.03749683380126953, 0.03755929565429687, 0.03760537719726562, 0.037743614196777346, 0.03752140808105469, 0.0375551986694336, 0.03762995147705078, 0.037495807647705076, 0.03751935958862305, 0.03787571334838867, 0.03770880126953125, 0.0376360969543457, 0.03753472137451172, 0.037599231719970705, 0.03759718322753906, 0.03856595230102539, 0.03801900863647461, 0.037601280212402347, 0.03762073516845703, 0.0376545295715332, 0.03752243041992188, 0.03753881454467774, 0.03760435104370117, 0.03775795364379883, 0.037550079345703126, 0.03786649703979492, 0.03831398391723633, 0.03769036865234375, 0.037601280212402347, 0.03755929565429687, 0.037498878479003905, 0.03746303939819336, 0.03752345657348633, 0.03794432067871094, 0.037713920593261716, 0.03760435104370117, 0.037495807647705076, 0.037591041564941405, 0.03748454284667969, 0.03763916778564453, 0.03788083267211914, 0.07681126403808594, 0.03763711929321289, 0.03831808090209961, 0.03748966217041016, 0.03763507080078125, 0.03758489608764649, 0.03759308624267578, 0.037579776763916016, 0.03772419357299805, 0.03755926513671875, 0.03763097763061524, 0.03747020721435547, 0.03756748962402344, 0.03761663818359375, 0.037901313781738284, 0.03767398452758789, 0.03793817520141601, 0.03772415924072266, 0.037601280212402347, 0.03751321411132812, 0.03760332870483398, 0.03770470428466797, 0.03760435104370117, 0.03767705535888672, 0.03810611343383789, 0.0376995849609375, 0.03758694458007812, 0.03785728073120117, 0.03757878494262695, 0.03765244674682617, 0.03786956787109375, 0.03756032180786133, 0.037838848114013675, 0.037773311614990236, 0.037525505065917966, 0.037884929656982425, 0.03775692749023438, 0.03761151885986328, 0.0378869743347168, 0.03772518539428711, 0.037797889709472655, 0.03774566268920899, 0.037759998321533206, 0.03785113525390625, 0.038040576934814455, 0.03770470428466797, 0.0377077751159668, 0.03755929565429687, 0.037599231719970705, 0.037661697387695314, 0.037553150177001955, 0.037823486328125, 0.03759820938110352, 0.03769651031494141, 0.037648384094238284, 0.03772313690185547, 0.03791667175292969, 0.03774259185791016, 0.03760435104370117, 0.03763711929321289, 0.03762483215332031, 0.03786444854736328, 0.03789516830444336, 0.07634636688232421, 0.03765350341796875, 0.03746713638305664, 0.03770470428466797, 0.03758796691894531, 0.03751116943359375, 0.03785318374633789, 0.037889022827148434, 0.03752345657348633, 0.0374774398803711, 0.037649345397949216, 0.03765760040283203, 0.037743614196777346, 0.0377149429321289, 0.03764329528808594, 0.03746198272705078, 0.037561344146728515, 0.037763072967529294, 0.03768832015991211, 0.03759823989868164, 0.037550048828125, 0.03745382308959961, 0.03772825622558594, 0.03768012619018555, 0.03773952102661133, 0.03763097763061524, 0.03759820938110352, 0.03765657424926758, 0.0381921272277832, 0.037905406951904294, 0.03770675277709961, 0.03752345657348633, 0.03779072189331055, 0.0376627197265625, 0.03760844802856445, 0.03748761749267578, 0.03757670211791992, 0.0375654411315918, 0.03750713729858399, 0.03755820846557617, 0.0375551986694336, 0.03763302230834961, 0.037645313262939455, 0.03746099090576172, 0.03763507080078125, 0.0376258544921875, 0.03769651031494141, 0.0375203857421875, 0.037610496520996094, 0.037612545013427735, 0.03768627166748047, 0.03968307113647461, 0.03810201644897461, 0.03744870376586914, 0.03748761749267578, 0.03760844802856445, 0.037705726623535156, 0.0375654411315918, 0.037533695220947266, 0.037482494354248046, 0.037713920593261716, 0.037579776763916016, 0.037495807647705076, 0.07634534454345702, 0.03747840118408203, 0.037424129486083986, 0.03769753646850586, 0.03747020721435547, 0.03760639953613281, 0.03746815872192383, 0.03759308624267578, 0.03759001541137695, 0.037582847595214845, 0.037572608947753904, 0.037525505065917966, 0.03752959823608398, 0.03775590515136719, 0.03758899307250976, 0.037459968566894535, 0.0377446403503418, 0.037495807647705076, 0.03755929565429687, 0.037438465118408204, 0.03756851196289063, 0.037602302551269534, 0.037515262603759765, 0.03751321411132812, 0.03760639953613281, 0.037541889190673826, 0.03759206390380859, 0.03758694458007812, 0.03766681671142578, 0.037548030853271484, 0.037512191772460936, 0.037582847595214845, 0.037493759155273435, 0.037449726104736326, 0.03746815872192383, 0.03757056045532227, 0.037823486328125, 0.037482494354248046, 0.037579776763916016, 0.03767091369628906, 0.03766886520385742, 0.03751424026489258, 0.03748147201538086, 0.037495807647705076, 0.03765555191040039, 0.03773132705688476, 0.03749273681640625, 0.03760639953613281, 0.03768217468261719, 0.037533695220947266, 0.037394432067871096, 0.03765350341796875, 0.03762278366088867, 0.03748966217041016, 0.03768729782104492, 0.0376514892578125, 0.03766985702514648, 0.03745280075073242, 0.037541889190673826, 0.03762073516845703, 0.03751731109619141, 0.037572608947753904, 0.03764223861694336, 0.07663206481933593, 0.037482494354248046, 0.03758694458007812, 0.03754598236083984, 0.03749785614013672, 0.03743641662597656, 0.03762790298461914, 0.03742105484008789, 0.03755929565429687, 0.037628929138183595, 0.03765350341796875, 0.03744768142700195, 0.03746918487548828, 0.03767500686645508, 0.03757670211791992, 0.03751222229003906, 0.03749884796142578, 0.03762995147705078, 0.03771596908569336, 0.03759718322753906, 0.03778252792358398, 0.03771596908569336, 0.03758182525634766, 0.03766886520385742, 0.0377262077331543, 0.03749683380126953, 0.0374917106628418, 0.037571582794189456, 0.03759513473510742, 0.03750912094116211, 0.037544960021972655, 0.0376627197265625, 0.03753574371337891, 0.037466110229492186, 0.03755212783813477, 0.03748044967651367, 0.03741798400878906, 0.03748761749267578, 0.03759414291381836, 0.03739030456542969, 0.03777740859985351, 0.03743129730224609, 0.037466110229492186, 0.037466110229492186, 0.03757056045532227, 0.03745177459716797, 0.03782451248168945, 0.03930624008178711, 0.03780198287963867, 0.037553150177001955, 0.037572608947753904, 0.037628929138183595, 0.03765350341796875, 0.03766579055786133, 0.03758796691894531, 0.03759718322753906, 0.03784396743774414, 0.038711296081542966, 0.03820236968994141, 0.037833728790283204, 0.03772931289672852, 0.03775075149536133, 0.03758796691894531, 0.07656038665771485, 0.03744153594970703, 0.037580799102783204, 0.03758489608764649, 0.03763507080078125, 0.03759308624267578, 0.03766681671142578, 0.0375470085144043, 0.0374835205078125, 0.03772313690185547, 0.037703678131103514, 0.037541889190673826, 0.0375551986694336, 0.037652511596679684, 0.037497825622558593, 0.03801190567016602, 0.037644287109375, 0.03753881454467774, 0.03759718322753906, 0.03752345657348633, 0.037541889190673826, 0.03757363128662109, 0.0374835205078125, 0.03762176132202148, 0.03808870315551758, 0.03783782577514649, 0.03767705535888672, 0.03754086303710937, 0.03700121688842774, 0.03764223861694336, 0.03752959823608398, 0.037765121459960936, 0.03765862274169922, 0.03768012619018555, 0.03785116958618164, 0.03752854537963867, 0.03746201705932617, 0.037591041564941405, 0.03764022445678711, 0.03770057678222656, 0.03760435104370117, 0.037569534301757815, 0.03757875061035156, 0.037550079345703126, 0.03768320083618164, 0.03784806442260742, 0.037806079864501956, 0.037579776763916016, 0.03758796691894531, 0.03765248107910156, 0.03765862274169922, 0.03774566268920899, 0.03764121627807617, 0.03772313690185547, 0.03784294509887695, 0.037817344665527344, 0.03783987045288086, 0.03761971282958984, 0.03763814544677734, 0.03765862274169922, 0.0377077751159668, 0.03765862274169922, 0.03762995147705078, 0.07627571105957032, 0.03767603302001953, 0.03746713638305664, 0.03762278366088867, 0.03759001541137695, 0.037689342498779296, 0.03778559875488281, 0.03764019012451172, 0.03788390350341797, 0.03751424026489258, 0.03749273681640625, 0.03770880126953125, 0.03754291152954101, 0.037703678131103514, 0.03769036865234375, 0.037574657440185545, 0.03785420989990235, 0.03765657424926758, 0.03755929565429687, 0.037599231719970705, 0.03741593551635742, 0.03765555191040039, 0.0375838737487793, 0.038059009552001956, 0.03752243041992188, 0.03781324768066406, 0.03762790298461914, 0.03782761764526367, 0.03764118576049805, 0.03759513473510742, 0.03786444854736328, 0.037617664337158206, 0.03751731109619141, 0.037566463470458986, 0.03778252792358398, 0.037596160888671876, 0.03751321411132812, 0.03806412887573242, 0.03936972808837891, 0.03786342239379883, 0.03775897598266602, 0.037601280212402347, 0.03762483215332031, 0.03776921463012695, 0.03782860946655273, 0.037449726104736326, 0.037705726623535156, 0.0375654411315918, 0.03755929565429687, 0.03753267288208008, 0.03763302230834961, 0.03753881454467774, 0.037580799102783204, 0.03768832015991211, 0.037574657440185545, 0.037531646728515625, 0.03764940643310547, 0.03754393768310547, 0.03761151885986328, 0.037515262603759765, 0.03761459350585938, 0.03788288116455078, 0.03977830505371094, 0.07646412658691407, 0.03738828659057617, 0.03792588806152344, 0.037748737335205076, 0.03746713638305664, 0.037897216796875, 0.037553150177001955, 0.03762176132202148, 0.037343231201171875, 0.037651454925537106, 0.037405696868896485, 0.0374835205078125, 0.03782758331298828, 0.03766681671142578, 0.03759001541137695, 0.0375838737487793, 0.0378081283569336, 0.03773235321044922, 0.03765760040283203, 0.03766886520385742, 0.03765964889526367, 0.03771187210083008, 0.03750707244873047, 0.03753472137451172, 0.037531646728515625, 0.03755110549926758, 0.03769343948364258, 0.03748659133911133, 0.037596160888671876, 0.03750707244873047, 0.03753881454467774, 0.037820415496826174, 0.03772415924072266, 0.03755929565429687, 0.037553150177001955, 0.03769036865234375, 0.03775897598266602, 0.037594112396240234, 0.03757875061035156, 0.03759513473510742, 0.03752755355834961, 0.03780505752563477, 0.03753267288208008, 0.03759820938110352, 0.03762790298461914, 0.03850447845458985, 0.03777225494384766, 0.03763916778564453, 0.03750297546386719, 0.037561344146728515, 0.037645313262939455, 0.03790848159790039, 0.037582847595214845, 0.037553150177001955, 0.037664768218994144, 0.03753574371337891, 0.03770675277709961, 0.03749075317382813, 0.03756332778930664, 0.03781324768066406, 0.03759718322753906, 0.037773311614990236, 0.037571582794189456, 0.07186124420166015, 0.037190654754638675, 0.03747225570678711, 0.03754908752441406, 0.037579742431640625, 0.03760639953613281, 0.03770982360839844, 0.03739136123657227, 0.03773440170288086, 0.0376360969543457, 0.03769241714477539, 0.03769241714477539, 0.03752345657348633, 0.03768627166748047, 0.03762688064575195, 0.03752140808105469, 0.037733375549316404, 0.03770675277709961, 0.037754878997802735, 0.03753472137451172, 0.03745792007446289, 0.03804876708984375, 0.03773747253417969, 0.037572608947753904, 0.03754291152954101, 0.03757056045532227, 0.03814297485351562, 0.03765350341796875, 0.03766988754272461, 0.03759001541137695, 0.03777536010742188, 0.03753472137451172, 0.037574657440185545, 0.03759718322753906, 0.03750707244873047, 0.03766886520385742, 0.03754393768310547, 0.03745280075073242, 0.03773952102661133, 0.03783168029785156, 0.03754086303710937, 0.03762688064575195, 0.037483551025390624, 0.03754902267456055, 0.03769036865234375, 0.03759718322753906, 0.037456897735595705, 0.03748761749267578, 0.038013950347900394, 0.037768192291259765, 0.03762073516845703, 0.03767705535888672, 0.037928958892822266, 0.03753779220581055, 0.03802828979492188, 0.037585918426513674, 0.039975936889648435, 0.03809791946411133, 0.03769343948364258, 0.037564414978027344, 0.03762278366088867, 0.03766988754272461, 0.03770470428466797, 0.07192781066894531, 0.03523993682861328, 0.03559731292724609, 0.038373374938964845, 0.03784908676147461, 0.03781017684936523, 0.03761663818359375, 0.037765121459960936, 0.037564414978027344, 0.037615615844726565, 0.03755417633056641, 0.038040576934814455, 0.03775692749023438, 0.03775692749023438, 0.037582847595214845, 0.037716991424560545, 0.03759308624267578, 0.03775283050537109, 0.03753062438964844, 0.03760435104370117, 0.03752447891235351, 0.037569534301757815, 0.03765964889526367, 0.03779891204833984, 0.03758182525634766, 0.037698558807373043, 0.037585918426513674, 0.03753062438964844, 0.03767398452758789, 0.037664768218994144, 0.0376627197265625, 0.037787647247314454, 0.03765657424926758, 0.037833728790283204, 0.037781505584716796, 0.037684223175048825, 0.037719039916992186, 0.03794944000244141, 0.037571582794189456, 0.03768627166748047, 0.037579776763916016, 0.037574657440185545, 0.03738521575927734, 0.03818188858032227, 0.0376627197265625, 0.03749990463256836, 0.037795841217041014, 0.03761151885986328, 0.037803009033203126, 0.0377077751159668, 0.03760435104370117, 0.03778867340087891, 0.037615615844726565, 0.03751731109619141, 0.03779379272460937, 0.03823820877075195, 0.03774054336547852, 0.03762688064575195, 0.03750400161743164, 0.037553150177001955, 0.037956607818603515, 0.03771289443969727, 0.037722110748291016]",tokens/s,26.191427819286087,, @@ -2232,7 +2232,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8b36-0cd0d3d161dd8e8c408b679f;c49ebf35-5ffe-4f73-9d99-3ac9686dbf6a) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e430-5468c04e726e2c74679a4228;0405280a-e9fc-4210-aa33-38635889f343) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2302,7 +2302,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f9014-379f446e4a508c602ae20df3;50dcaf06-ce11-4f65-9469-b0df3e38cd38) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e8d5-70e89f185727510c73738967;3cc09ca1-ae96-460e-996e-5b8cf90dc326) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2372,7 +2372,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8fbe-471219ba308e4c0c2b78317e;35bd23e8-7b41-406f-a3d5-3e443fae0702) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e872-08abaf584811ad3f0149a204;89237cb0-6278-4e68-91eb-733c953f957a) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2472,7 +2472,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8e7c-2211d56c710cc0bc797963c4;1384795c-94b1-4e13-9f45-dfd1bec95807) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e730-459ae65c7fba894471a43ff1;95fec30d-4670-4064-b091-0e2ac3e59557) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2571,7 +2571,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8f00-5f2c4c677f282a8342807129;f649dc00-f657-4a16-b325-b59a5eb91555) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e7a9-7c43db445544a3ef296e8968;dd9cad12-58b5-4630-9bfb-c5f87fc1223a) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2839,7 +2839,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666f87fb-2532bf75427d3adc2fed5ffd;845b4bd3-1b89-4ec8-8deb-ba7eefcb0551) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d801-77baef2c69b0a7b721ddb518;ef0f1c61-49db-4c8f-ae37-2cd8992f2ac9) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -3044,7 +3044,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8ea6-2fddb0a94c0f756558e3244a;039949f1-7bb4-46ca-a705-4b6b18118282) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e759-35c37dd32d6067913dbfad81;58e5bb18-baf8-45da-b42b-464d971006cb) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -3104,7 +3104,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666f87c9-04796c2f0d62315a54ec91a3;549d3507-eca0-4b37-ba1c-912c9bc045d0) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d7d8-41f1c07828feda1e49441ab2;422d2d3f-607d-4023-89cc-41c914d48a53) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -3629,7 +3629,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666f8536-30a0a70a04c6c5cc24d1997e;e03ff209-55ab-46aa-a36f-31ea0b361057) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d63f-70f76ecd412f5301068afe41;209659a7-e8dc-4734-ba46-8a3ce588de12) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -3714,7 +3714,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8e24-662b0ab9328447db2ad2818b;efc39137-2367-4cd1-8f18-f4f4fc3b7f6a) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e6e0-5bee696633c35abb10fa0a95;9727e1cb-6a9e-485f-a0c7-d7f9631049d8) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -3858,7 +3858,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8d2d-3c1209ad4c8df5e243fa34a9;cdc51707-5d73-4a76-9839-caf557ac38a4) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e617-0b5658260d07187538508fdb;6d985839-5386-4cfb-8971-763cde79472e) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4030,7 +4030,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8f84-36e4040c57710f7c537f32ed;657c4c68-4773-49b6-bee0-94a24df3d3f4) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e82e-25aa29b72542f13b579f7793;091c0572-a6d5-4097-8654-2f98d056d8d5) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4129,7 +4129,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8cfe-4c2eb36a46f66c3414114ae3;f26edc09-6e65-4d74-b4d0-4e1fc1fa8b4c) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e5ee-1f6a30f35289df7832a44b68;a953ad48-e602-449a-90b7-d72db92354f0) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4228,7 +4228,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8dc9-6e1d80462e9a20193a2f2e1d;d87a1029-82b0-4e43-b01f-ff09ab9cba90) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e68f-3e23550579095d4754821fb8;40673687-a4b7-48ec-a9ff-0f499f09ba4d) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4319,7 +4319,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666f913e-105b75d821f2492927206f1e;54959402-f92a-41b0-ab9d-116a54f948e8) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e9d5-712824e953f911ef7e5c58ca;6968df34-8595-4dc4-ba9c-1dfac596ec7d) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -4404,7 +4404,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8df8-0f2c56f6671703777d206010;7e860c7e-563e-4e56-a77a-0e3329130b52) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e6b7-14b3df764cc7dd967d2456d9;32b0290b-8046-436d-8293-0a4440113c59) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4494,7 +4494,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666f8793-21b223517d76255702dbc8dd;ca20dedd-333d-462b-bd6a-8ed96b2e55e6) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d7b1-12f8e71a475ed55d697c7d62;c9fa4803-9eb4-4367-ab6c-5f9e46b54d1c) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -4677,7 +4677,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8d9d-5e00ac5701536b8f661823eb;b99e147f-8c2c-4498-9955-0557b7178ca7) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e667-77e7adae7899b3d4220679d8;296772ba-550f-4fa1-844b-f2e67a133a88) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4806,7 +4806,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8824-1e3d706f1f93519055bf870c;c6f5cae5-27cf-4c11-b0fe-5a97f581cefd) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670d829-3b52238a3e27ab8f400fd2b6;f9ed387b-3879-42f2-82a4-12c5f57f0364) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4935,7 +4935,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8f2e-32a2d6313912494141e5400e;1d576b02-7e87-4eb8-8c36-dc48ebe4a95e) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e7d1-1e05ac6950a6170949aaf0e1;2c62c82c-8a82-4207-9a9e-83446e50f4f1) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5035,7 +5035,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8b7f-3437a2ad7f1e3db95373385a;661c43cd-29d0-4087-89e1-7637b9cd870c) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e47f-28f7df0f51bf57b4324c6ede;8e34a220-ee5f-4845-942d-2e3841445641) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5344,39 +5344,7 @@ ValueError: OPTForCausalLM does not support an attention implementation through ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,microsoft/phi-1_5,microsoft/phi-1_5,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1407.352832,2078.80192,0.0,1432.354816,1207.615488,s,10,0.6952154235839844,0.06952154235839844,0.0003284714911214616,0.06937214279174805,0.06993558425903319,0.07001076850891114,0.07007091590881348,"[0.06991887664794921, 0.06913164520263672, 0.06926902770996093, 0.0692248306274414, 0.06923782348632812, 0.06943878173828125, 0.06930550384521485, 0.06973856353759765, 0.07008595275878907, 0.06986441802978516]",tokens/s,3682.31186069298,kWh,8.181360695097182e-07,4.4830105463790514e-07,3.5595967174070095e-06,4.826033841554632e-06,tokens/kWh,53045628.85483902,MB,1407.352832,2078.80192,0.0,1432.354816,1306.320384,s,10,36.407312255859374,3.640731225585937,0.0010796254011809284,3.6406092529296874,3.641276708984375,3.64250322265625,3.64348443359375,"[3.643729736328125, 3.640717529296875, 3.639988525390625, 3.64072265625, 3.641004150390625, 3.64038671875, 3.640628662109375, 3.63993017578125, 3.6396142578125, 3.64058984375]",tokens/s,17.304216130335416,kWh,4.2991543031401105e-05,2.3561688259343446e-05,0.00018775002855239066,0.0002543032598431352,tokens/kWh,247735.71537722726,,s,629,36.97445987701416,0.05878292508269341,0.008278403521448836,0.05777510452270508,0.057852314758300785,0.05792010269165039,0.1272576400756836,"[0.058324993133544924, 0.05831679916381836, 0.05833216094970703, 0.05831372833251953, 0.058298366546630856, 0.05830863952636719, 0.05800342559814453, 0.058352638244628906, 0.057815040588378906, 0.05774028778076172, 0.057761791229248044, 0.05778124618530273, 0.05775667190551758, 0.057761791229248044, 0.05774028778076172, 0.057745407104492184, 0.05774131011962891, 0.05773004913330078, 0.057777153015136716, 0.05776588821411133, 0.05776793670654297, 0.05766758346557617, 0.05774335861206055, 0.05770035171508789, 0.057761791229248044, 0.05777305603027344, 0.05769830322265625, 0.0577525749206543, 0.05770035171508789, 0.057768959045410156, 0.057763839721679686, 0.05781708908081055, 0.05770137786865234, 0.057734142303466796, 0.05775155258178711, 0.057778175354003904, 0.05782527923583984, 0.05775872039794922, 0.057745407104492184, 0.057780223846435545, 0.05775462341308594, 0.05780684661865235, 0.05774028778076172, 0.05781094360351562, 0.057777153015136716, 0.057785343170166016, 0.05778841781616211, 0.05784473419189453, 0.05779046249389649, 0.057717758178710936, 0.05770035171508789, 0.057801727294921876, 0.057768959045410156, 0.05771878433227539, 0.057763839721679686, 0.057731071472167966, 0.057809921264648435, 0.0576808967590332, 0.0577710075378418, 0.057763839721679686, 0.05777510452270508, 0.05774131011962891, 0.12719615936279297, 0.05770751953125, 0.05772185516357422, 0.05774643325805664, 0.057799678802490234, 0.05775360107421875, 0.05772697448730469, 0.05781094360351562, 0.05770035171508789, 0.05778432083129883, 0.057745407104492184, 0.05773311996459961, 0.05773926544189453, 0.05788979339599609, 0.057796607971191405, 0.057780223846435545, 0.057695232391357425, 0.05787648010253906, 0.057785343170166016, 0.05776998519897461, 0.05795635223388672, 0.05787443161010742, 0.05772288131713867, 0.057760768890380856, 0.05763993453979492, 0.05777407836914063, 0.05775155258178711, 0.057731071472167966, 0.057624576568603515, 0.05773619079589844, 0.05777612686157227, 0.05776998519897461, 0.05772390365600586, 0.05787340927124023, 0.057783294677734375, 0.05771263885498047, 0.05767679977416992, 0.05783244705200195, 0.05783859252929688, 0.05774131011962891, 0.057717758178710936, 0.05782220840454102, 0.05779558563232422, 0.05780684661865235, 0.05773516845703125, 0.05778739166259766, 0.057823230743408206, 0.057783294677734375, 0.05780582427978516, 0.05792768096923828, 0.05790412902832031, 0.05773311996459961, 0.057763839721679686, 0.0580761604309082, 0.057780223846435545, 0.057801727294921876, 0.05770547103881836, 0.05782425689697265, 0.05773516845703125, 0.05777203369140625, 0.057708545684814455, 0.05785190582275391, 0.058226688385009766, 0.1272668151855469, 0.05778739166259766, 0.05778227233886719, 0.05767679977416992, 0.05777920150756836, 0.057750526428222655, 0.057763839721679686, 0.057766910552978515, 0.05773619079589844, 0.05811814498901367, 0.05778739166259766, 0.05778227233886719, 0.05768806457519531, 0.05784064102172851, 0.05771673583984375, 0.05779046249389649, 0.05778739166259766, 0.05778124618530273, 0.057919486999511716, 0.05782527923583984, 0.057744384765625, 0.05773823928833008, 0.057799678802490234, 0.057785343170166016, 0.05774233627319336, 0.05770240020751953, 0.05780684661865235, 0.05770342254638672, 0.05774335861206055, 0.0576808967590332, 0.05777920150756836, 0.05782425689697265, 0.05781913757324219, 0.057734142303466796, 0.05774643325805664, 0.057706497192382813, 0.05778739166259766, 0.0577894401550293, 0.0577525749206543, 0.05772288131713867, 0.05777203369140625, 0.057706497192382813, 0.057793537139892576, 0.057708545684814455, 0.057839614868164066, 0.05776588821411133, 0.05784473419189453, 0.05780275344848633, 0.05780889511108398, 0.05779865646362305, 0.05774950408935547, 0.057760768890380856, 0.057778175354003904, 0.05784678268432617, 0.057711616516113284, 0.05778636932373047, 0.057761791229248044, 0.05779251098632812, 0.05770956802368164, 0.05780889511108398, 0.057791488647460934, 0.05777510452270508, 0.05773004913330078, 0.12723404693603516, 0.05772185516357422, 0.057717758178710936, 0.05775872039794922, 0.05778227233886719, 0.057785343170166016, 0.05780889511108398, 0.05777203369140625, 0.05773311996459961, 0.057768959045410156, 0.05779251098632812, 0.05776998519897461, 0.057750526428222655, 0.05781196975708008, 0.057791488647460934, 0.057768959045410156, 0.057708545684814455, 0.05779558563232422, 0.057761791229248044, 0.057793537139892576, 0.057747455596923826, 0.057807872772216794, 0.05775974273681641, 0.0577710075378418, 0.05768499374389648, 0.05782425689697265, 0.05778227233886719, 0.057728000640869144, 0.057706497192382813, 0.057777153015136716, 0.05779763031005859, 0.05776588821411133, 0.05775769424438477, 0.05771468734741211, 0.05777305603027344, 0.05774233627319336, 0.05777407836914063, 0.05776588821411133, 0.05779558563232422, 0.05776486587524414, 0.05775462341308594, 0.05777920150756836, 0.05788774490356445, 0.05777920150756836, 0.05789286422729492, 0.05776588821411133, 0.05784678268432617, 0.05780684661865235, 0.05786316680908203, 0.05780582427978516, 0.05789798355102539, 0.057760768890380856, 0.0577710075378418, 0.05782425689697265, 0.0577525749206543, 0.057919486999511716, 0.0577525749206543, 0.05778636932373047, 0.057875457763671874, 0.05784985733032227, 0.057785343170166016, 0.057903102874755856, 0.05783039855957031, 0.12727808380126954, 0.05815500640869141, 0.057862144470214844, 0.0577259521484375, 0.05777510452270508, 0.05776793670654297, 0.057815040588378906, 0.05778739166259766, 0.057706497192382813, 0.057717758178710936, 0.05778739166259766, 0.05776588821411133, 0.05781094360351562, 0.05805670547485352, 0.05772697448730469, 0.05777920150756836, 0.0577259521484375, 0.0578682861328125, 0.05781401443481445, 0.057801727294921876, 0.057711616516113284, 0.057796607971191405, 0.05771980667114258, 0.057815040588378906, 0.05770137786865234, 0.057777153015136716, 0.057799678802490234, 0.057662464141845705, 0.0576890869140625, 0.057785343170166016, 0.05777305603027344, 0.05781708908081055, 0.05782015991210938, 0.05772288131713867, 0.057809921264648435, 0.05783654403686524, 0.05779455947875976, 0.057768959045410156, 0.05780889511108398, 0.05774335861206055, 0.05779865646362305, 0.0577628173828125, 0.05785497665405273, 0.057801727294921876, 0.05781913757324219, 0.05773209762573242, 0.05784064102172851, 0.057829376220703124, 0.05779455947875976, 0.05779558563232422, 0.05780889511108398, 0.05779251098632812, 0.05779558563232422, 0.057761791229248044, 0.05775155258178711, 0.057842689514160155, 0.05779455947875976, 0.05782425689697265, 0.057708545684814455, 0.05781401443481445, 0.05778636932373047, 0.05775462341308594, 0.05775564956665039, 0.12743270111083985, 0.05773619079589844, 0.057812992095947265, 0.057785343170166016, 0.05780377578735352, 0.05777407836914063, 0.05778432083129883, 0.05782220840454102, 0.0577259521484375, 0.05772288131713867, 0.057785343170166016, 0.05774950408935547, 0.05775360107421875, 0.05777407836914063, 0.057766910552978515, 0.057768959045410156, 0.057778175354003904, 0.05780377578735352, 0.057837566375732424, 0.05777510452270508, 0.05792051315307617, 0.05788876724243164, 0.05775564956665039, 0.05784883117675781, 0.05769420623779297, 0.057766910552978515, 0.057766910552978515, 0.05773823928833008, 0.05769830322265625, 0.05773004913330078, 0.05782425689697265, 0.05782835388183594, 0.05778636932373047, 0.05772185516357422, 0.057826305389404295, 0.057731071472167966, 0.05774028778076172, 0.05775462341308594, 0.057799678802490234, 0.057809921264648435, 0.05775564956665039, 0.05773209762573242, 0.05782527923583984, 0.05776486587524414, 0.05778841781616211, 0.057761791229248044, 0.05780582427978516, 0.057778175354003904, 0.05789286422729492, 0.057804798126220705, 0.057799678802490234, 0.05773926544189453, 0.057809921264648435, 0.057777153015136716, 0.05781708908081055, 0.05780070495605469, 0.057717758178710936, 0.057780223846435545, 0.057763839721679686, 0.05781913757324219, 0.05770956802368164, 0.057815040588378906, 0.05778124618530273, 0.12731597137451173, 0.057747455596923826, 0.05775769424438477, 0.05773004913330078, 0.057799678802490234, 0.057780223846435545, 0.057768959045410156, 0.057823230743408206, 0.05768703842163086, 0.05782220840454102, 0.057760768890380856, 0.05786316680908203, 0.05771980667114258, 0.05784371185302734, 0.05776588821411133, 0.057804798126220705, 0.05774950408935547, 0.05783347320556641, 0.057745407104492184, 0.057761791229248044, 0.05771366500854492, 0.05785599899291992, 0.05774950408935547, 0.05777305603027344, 0.05776998519897461, 0.057777153015136716, 0.057780223846435545, 0.05770547103881836, 0.05775155258178711, 0.05773823928833008, 0.057778175354003904, 0.05779251098632812, 0.057829376220703124, 0.05772083282470703, 0.05772185516357422, 0.05774233627319336, 0.057793537139892576, 0.057777153015136716, 0.057804798126220705, 0.05785804748535156, 0.057801727294921876, 0.05774028778076172, 0.05783859252929688, 0.05779046249389649, 0.057778175354003904, 0.05772185516357422, 0.05807513427734375, 0.05788671875, 0.057823230743408206, 0.05788467025756836, 0.05785599899291992, 0.0577628173828125, 0.05781401443481445, 0.057766910552978515, 0.05775667190551758, 0.057780223846435545, 0.05774131011962891, 0.05777612686157227, 0.05780377578735352, 0.05777407836914063, 0.05776998519897461, 0.057778175354003904, 0.057734142303466796, 0.1277491226196289, 0.05774848175048828, 0.05778432083129883, 0.05772083282470703, 0.057815040588378906, 0.05775974273681641, 0.057831424713134766, 0.05781401443481445, 0.05774643325805664, 0.05777612686157227, 0.057823230743408206, 0.057750526428222655, 0.057750526428222655, 0.05777305603027344, 0.05818675231933594, 0.05780582427978516, 0.05775872039794922, 0.05783244705200195, 0.057809921264648435, 0.057768959045410156, 0.05774643325805664, 0.05775769424438477, 0.05771571350097656, 0.057875457763671874, 0.05769830322265625, 0.05771571350097656, 0.05775155258178711, 0.05772390365600586, 0.05769318389892578, 0.05772697448730469, 0.057823230743408206, 0.05774950408935547, 0.05778636932373047, 0.05773004913330078, 0.0577894401550293, 0.05770035171508789, 0.057763839721679686, 0.057724929809570315, 0.057815040588378906, 0.05773209762573242, 0.057766910552978515, 0.05769830322265625, 0.05782425689697265, 0.05770956802368164, 0.05777920150756836, 0.05773823928833008, 0.05779763031005859, 0.05776588821411133, 0.05780684661865235, 0.057804798126220705, 0.05781913757324219, 0.05771059036254883, 0.057793537139892576, 0.05778124618530273, 0.057734142303466796, 0.05775564956665039, 0.05770444869995117, 0.057880577087402345, 0.057763839721679686, 0.05781708908081055, 0.05771059036254883, 0.05777920150756836, 0.05773516845703125, 0.12806246948242186, 0.0577259521484375, 0.05774131011962891, 0.05770240020751953, 0.0577628173828125, 0.05774848175048828, 0.05780275344848633, 0.05777407836914063, 0.0577259521484375, 0.05775974273681641, 0.05778227233886719, 0.057761791229248044, 0.0576993293762207, 0.05781401443481445, 0.057812992095947265, 0.0577710075378418, 0.057673728942871094, 0.05790924835205078, 0.05778124618530273, 0.05777203369140625, 0.057678848266601565, 0.057799678802490234, 0.057744384765625, 0.05775462341308594, 0.05768396759033203, 0.05776486587524414, 0.05776588821411133, 0.0577894401550293, 0.057665534973144535, 0.057799678802490234, 0.057761791229248044, 0.057763839721679686, 0.0577259521484375, 0.057780223846435545, 0.057734142303466796, 0.05772083282470703, 0.05794713592529297, 0.057809921264648435, 0.05783039855957031, 0.057780223846435545, 0.05774950408935547, 0.05780275344848633, 0.05784371185302734, 0.057731071472167966, 0.05773823928833008, 0.057780223846435545, 0.057847808837890625, 0.057821182250976565, 0.05772083282470703, 0.05792153549194336, 0.05777920150756836, 0.0577628173828125, 0.05772185516357422, 0.05780070495605469, 0.057724929809570315, 0.05777612686157227, 0.05766758346557617, 0.05783552169799805, 0.057777153015136716, 0.05775769424438477, 0.05769830322265625, 0.05784371185302734, 0.05774131011962891, 0.12786380767822267, 0.05776793670654297, 0.057750526428222655, 0.05771980667114258, 0.057785343170166016, 0.05774233627319336, 0.05772390365600586, 0.05773619079589844, 0.05768499374389648, 0.057724929809570315, 0.05775564956665039, 0.05774848175048828, 0.05771366500854492, 0.057845760345458984, 0.05772185516357422, 0.057747455596923826, 0.05775564956665039, 0.05777612686157227, 0.057834495544433595, 0.05780275344848633, 0.05776793670654297, 0.057724929809570315, 0.05775564956665039, 0.057777153015136716, 0.05770342254638672, 0.05768806457519531, 0.05780070495605469, 0.057870334625244144, 0.057780223846435545, 0.057675777435302736, 0.057778175354003904, 0.057785343170166016, 0.05775667190551758, 0.057706497192382813, 0.05775155258178711, 0.057734142303466796, 0.05780377578735352, 0.05780377578735352, 0.05811507034301758, 0.05779046249389649, 0.05776486587524414, 0.05776588821411133, 0.05777407836914063, 0.05789491271972656, 0.05782527923583984, 0.05788979339599609, 0.057809921264648435, 0.05782220840454102, 0.05795532989501953, 0.057815040588378906, 0.05778124618530273, 0.05777203369140625, 0.05778432083129883, 0.05783039855957031, 0.057853950500488284, 0.057826305389404295, 0.057796607971191405, 0.05778124618530273, 0.05770547103881836, 0.057940990447998046, 0.05777920150756836, 0.05782527923583984, 0.057717758178710936]",tokens/s,17.01174275681656,, -4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,internlm/internlm2-20b,internlm/internlm2-20b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch - benchmark_report = Benchmark.launch(benchmark_config) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch - report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch - raise ChildProcessError(response[""traceback""]) -ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target - report = worker(*worker_args) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run - backend: Backend = backend_factory(backend_config) - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 79, in __init__ - self.load_model_with_no_weights() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 255, in load_model_with_no_weights - self.load_model_from_pretrained() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 169, in load_model_from_pretrained - self.pretrained_model = self.automodel_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 558, in from_pretrained - return model_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3626, in from_pretrained - model = cls(config, *model_args, **model_kwargs) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/f363ea8a116b3ea829c7a068ca24bc9d3e668083/modeling_internlm2.py"", line 974, in __init__ - self.model = InternLM2Model(config) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/f363ea8a116b3ea829c7a068ca24bc9d3e668083/modeling_internlm2.py"", line 796, in __init__ - self.layers = nn.ModuleList([InternLM2DecoderLayer(config) for _ in range(config.num_hidden_layers)]) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/f363ea8a116b3ea829c7a068ca24bc9d3e668083/modeling_internlm2.py"", line 796, in - self.layers = nn.ModuleList([InternLM2DecoderLayer(config) for _ in range(config.num_hidden_layers)]) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/f363ea8a116b3ea829c7a068ca24bc9d3e668083/modeling_internlm2.py"", line 598, in __init__ - self.attention = INTERNLM2_ATTENTION_CLASSES[config.attn_implementation](config=config) -KeyError: 'sdpa' - -",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,internlm/internlm2-20b,internlm/internlm2-20b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,,,,,MB,3137.929216,13793.492992,0.0,13147.045888,12781.924864,s,10,8.99033642578125,0.8990336425781249,0.001209152583374722,0.8988238220214844,0.9007634643554687,0.9011640808105469,0.9014845739746095,"[0.898102294921875, 0.8976475830078126, 0.89875732421875, 0.8988903198242187, 0.89919384765625, 0.8981031494140626, 0.8995744018554688, 0.9006744384765625, 0.901564697265625, 0.897828369140625]",tokens/s,284.7501893987843,kWh,1.0606409098815037e-05,5.811931839480167e-06,5.64317581083322e-05,7.285009904662741e-05,tokens/kWh,3514065.229151552,MB,3137.929216,13793.492992,0.0,13147.045888,12906.706432,s,10,495.9275859375,49.59275859375,0.0018491618139201286,49.592791015625,49.594271093749995,49.595561328125,49.596593515625,"[49.5933984375, 49.590046875, 49.59246875, 49.5923359375, 49.593984375, 49.59295703125, 49.5899609375, 49.592640625, 49.5968515625, 49.59294140625]",tokens/s,1.2703467559866628,kWh,0.0005854946296717283,0.00032090146071840533,0.0030803726957776686,0.003986768786167803,tokens/kWh,15802.270806017175,,s,629,503.2330070800783,0.8000524754850208,0.10683804867936002,0.78717236328125,0.7874144165039062,0.7875112915039062,1.685771015625,"[0.78681396484375, 0.787056640625, 0.7872000122070313, 0.7871314086914063, 0.7870761108398437, 0.7871968994140625, 0.7870853271484375, 0.7871027221679687, 0.7870504760742187, 0.7872420043945313, 0.7869163818359375, 0.7871047973632812, 0.7869481201171875, 0.7870822143554688, 0.787114990234375, 0.7870955810546875, 0.78725634765625, 0.7873034057617188, 0.7869389038085938, 0.787019775390625, 0.7870382080078125, 0.7876044921875, 0.7871713256835937, 0.7871876831054687, 0.7872317504882812, 0.7870443725585937, 0.787030029296875, 0.7872450561523437, 0.7871856689453125, 0.7870392456054688, 0.7871764526367188, 0.7872102661132813, 0.7872501831054688, 0.7873095703125, 0.7874119873046875, 0.787078125, 0.7872286987304687, 0.7874293823242188, 0.787230712890625, 0.7871242065429688, 0.7872471313476562, 0.7872798461914062, 0.7871795043945312, 0.7870382080078125, 0.7872808837890625, 0.7874662475585937, 0.7871713256835937, 0.7872593994140625, 0.7871129760742187, 0.7872962646484375, 0.7874160766601562, 0.7872471313476562, 0.7874232177734375, 0.787145751953125, 0.78729931640625, 0.7873556518554687, 0.7873126220703125, 0.7872225341796875, 0.7873648681640625, 0.7873320922851562, 0.7873925170898437, 0.787409912109375, 1.6845506591796875, 0.7865989379882813, 0.7869122314453125, 0.7870771484375, 0.786904052734375, 0.786988037109375, 0.7868211059570313, 0.7866941528320313, 0.7869010009765625, 0.7872276611328125, 0.7870422973632812, 0.7869706420898438, 0.78719384765625, 0.786861083984375, 0.78710888671875, 0.7869931640625, 0.7870863647460937, 0.7872379150390625, 0.7870750732421875, 0.786935791015625, 0.7871508178710938, 0.7870403442382813, 0.7873442993164063, 0.7869214477539063, 0.7870392456054688, 0.7870474243164063, 0.786840576171875, 0.7870504760742187, 0.78713037109375, 0.7870873413085937, 0.7869808349609375, 0.7869808349609375, 0.787267578125, 0.786956298828125, 0.787124267578125, 0.7870125732421875, 0.786892822265625, 0.7871385498046874, 0.78703515625, 0.7871498413085938, 0.7871344604492188, 0.78739453125, 0.7875164184570312, 0.7874877319335938, 0.7870628051757812, 0.787272705078125, 0.787435546875, 0.7874375610351563, 0.787373046875, 0.7877529296875, 0.7874273071289063, 0.787282958984375, 0.7873843383789062, 0.7874119873046875, 0.7875717163085938, 0.7875112915039062, 0.7872808837890625, 0.7871426391601563, 0.7872798461914062, 0.7872686157226563, 0.7872399291992187, 0.7873178100585938, 0.78739453125, 1.68635595703125, 0.7868969116210938, 0.787056640625, 0.7871334228515625, 0.7868477172851562, 0.7871549682617187, 0.787078125, 0.7870904541015625, 0.7870392456054688, 0.787114990234375, 0.7871692504882812, 0.7877386474609375, 0.7869378662109375, 0.7871314086914063, 0.7872255859375, 0.7870453491210937, 0.7869655151367188, 0.786951171875, 0.7870986328125, 0.7870658569335938, 0.7872614135742187, 0.7874027709960938, 0.7870945434570312, 0.78715185546875, 0.7872450561523437, 0.7871549682617187, 0.7872819213867187, 0.787478515625, 0.7873597412109375, 0.7869859619140624, 0.7874692993164063, 0.7872286987304687, 0.7869798583984375, 0.787325927734375, 0.7869276123046876, 0.7874866943359375, 0.7870545654296875, 0.787198974609375, 0.7871416015625, 0.787103759765625, 0.7871918334960938, 0.787167236328125, 0.7870433349609375, 0.78729931640625, 0.7875491943359375, 0.7871170654296875, 0.787009521484375, 0.7875952758789062, 0.7872061157226562, 0.7870023803710937, 0.7870484619140625, 0.7871426391601563, 0.787188720703125, 0.7870167236328125, 0.7873402709960937, 0.7872317504882812, 0.7870904541015625, 0.7873126220703125, 0.7873116455078125, 0.78721533203125, 0.7872767944335938, 0.7872000122070313, 0.7872696533203125, 1.686043701171875, 0.7870218505859375, 0.7873935546875, 0.7871447143554687, 0.787346435546875, 0.7868549194335938, 0.7871016845703125, 0.7867627563476562, 0.7869368286132813, 0.7868671875, 0.7868692626953125, 0.786977783203125, 0.7869685668945312, 0.7871201171875, 0.7870084838867187, 0.7870965576171876, 0.7871498413085938, 0.7871580200195313, 0.7870658569335938, 0.7869450073242188, 0.7869378662109375, 0.787483642578125, 0.7873648681640625, 0.7870310668945313, 0.7869224853515625, 0.7871365966796875, 0.787136474609375, 0.7871539306640625, 0.787125244140625, 0.7869398803710937, 0.78703515625, 0.7872276611328125, 0.7871282958984375, 0.7870628051757812, 0.7872921752929688, 0.7870003051757812, 0.7870494995117188, 0.7871273193359375, 0.7872399291992187, 0.787398681640625, 0.78772119140625, 0.78713037109375, 0.787267578125, 0.7872276611328125, 0.7871488037109375, 0.787314697265625, 0.7874241943359375, 0.7870822143554688, 0.7873843383789062, 0.7872819213867187, 0.7872849731445313, 0.78731982421875, 0.7872379150390625, 0.787399658203125, 0.7873013916015625, 0.7877857055664063, 0.7872245483398438, 0.7873157348632812, 0.7872655639648437, 0.7875112915039062, 0.7873760986328125, 0.787336181640625, 0.78727783203125, 1.68654541015625, 0.7870822143554688, 0.7871160278320313, 0.7870023803710937, 0.7870228271484375, 0.7869972534179688, 0.786967529296875, 0.7874447631835938, 0.7872890625, 0.7869368286132813, 0.7870853271484375, 0.7870689086914062, 0.7870709838867187, 0.78703515625, 0.787505126953125, 0.7871375122070312, 0.7872081909179688, 0.7871528930664062, 0.7871610717773437, 0.787219482421875, 0.7872010498046875, 0.7872266235351563, 0.787040283203125, 0.7871918334960938, 0.7872061157226562, 0.7872471313476562, 0.7872041015625, 0.787078125, 0.7871713256835937, 0.7871109008789062, 0.7871968994140625, 0.7870965576171876, 0.7871549682617187, 0.78739453125, 0.787336181640625, 0.7870084838867187, 0.7874140014648437, 0.7870863647460937, 0.7871047973632812, 0.7871129760742187, 0.7871273193359375, 0.78738330078125, 0.7872819213867187, 0.7871928100585938, 0.7872757568359375, 0.7873280029296875, 0.7872593994140625, 0.7870822143554688, 0.7872921752929688, 0.7870812377929688, 0.78708837890625, 0.7870607299804687, 0.7874990234375, 0.7872245483398438, 0.78721435546875, 0.7873689575195313, 0.7873966064453125, 0.7875502319335937, 0.7875901489257813, 0.7873382568359375, 0.7871918334960938, 0.7875266723632812, 0.7872737426757812, 1.68506982421875, 0.7869859619140624, 0.7871129760742187, 0.7870054321289063, 0.7870218505859375, 0.78700341796875, 0.787030029296875, 0.7872501831054688, 0.7870105590820312, 0.786934814453125, 0.7869972534179688, 0.786946044921875, 0.787056640625, 0.787325927734375, 0.787252197265625, 0.7869522094726562, 0.7875215454101563, 0.7873843383789062, 0.78717236328125, 0.7871324462890625, 0.7873607788085938, 0.78706689453125, 0.7871395874023438, 0.78725634765625, 0.78721435546875, 0.7872757568359375, 0.7873546142578125, 0.7874447631835938, 0.7869368286132813, 0.78706689453125, 0.7870453491210937, 0.7871375122070312, 0.7872071533203125, 0.7872235717773437, 0.7871477661132813, 0.7870607299804687, 0.78715185546875, 0.7871478271484375, 0.7873289794921875, 0.7873003540039063, 0.787083251953125, 0.7871488037109375, 0.78725732421875, 0.7871795043945312, 0.78721533203125, 0.7876454467773437, 0.787178466796875, 0.7871815795898438, 0.7872112426757812, 0.78717236328125, 0.7872327880859376, 0.7870750732421875, 0.7871897583007812, 0.7872420043945313, 0.787294189453125, 0.7872420043945313, 0.7872491455078126, 0.7874774780273438, 0.7874058227539062, 0.7875645141601563, 0.7871815795898438, 0.7874334716796875, 0.7871918334960938, 1.687065673828125, 0.7868405151367187, 0.7869450073242188, 0.7869102172851562, 0.7868426513671875, 0.7869398803710937, 0.7869041137695313, 0.7869142456054687, 0.786946044921875, 0.786820068359375, 0.78691943359375, 0.787078125, 0.7869788208007813, 0.786951171875, 0.7870924682617187, 0.78710986328125, 0.7872061157226562, 0.787040283203125, 0.7869204711914063, 0.7873966064453125, 0.7871181030273438, 0.7870392456054688, 0.7870648193359375, 0.787114013671875, 0.7872133178710937, 0.7872122802734375, 0.7869757690429687, 0.7872870483398438, 0.7871006469726562, 0.7871600341796875, 0.7870965576171876, 0.7870996704101563, 0.7871242065429688, 0.7870771484375, 0.7871447143554687, 0.7871078491210938, 0.7870576782226563, 0.7873760986328125, 0.7871334228515625, 0.7871221923828124, 0.7870075073242188, 0.7875072021484375, 0.7872061157226562, 0.7871661987304688, 0.7876239624023438, 0.787252197265625, 0.7872225341796875, 0.7872481079101562, 0.78718359375, 0.7872593994140625, 0.7872921752929688, 0.7870996704101563, 0.7873925170898437, 0.78729931640625, 0.7872388916015625, 0.7874826049804687, 0.7871661987304688, 0.78746728515625, 0.7872122802734375, 0.7872655639648437, 0.787220458984375, 0.787262451171875, 0.7870750732421875, 1.6877587890625, 0.7871488037109375, 0.7872276611328125, 0.787061767578125, 0.7871856689453125, 0.7868661499023437, 0.7870802001953126, 0.7870126342773438, 0.787072998046875, 0.7868098754882813, 0.7869603881835937, 0.7869419555664062, 0.7873136596679687, 0.78706689453125, 0.787056640625, 0.7869767456054687, 0.7870965576171876, 0.7870637817382813, 0.7873054809570312, 0.7870341186523437, 0.7869490966796875, 0.7870863647460937, 0.7871129760742187, 0.78717236328125, 0.7871334228515625, 0.7870802001953126, 0.7873003540039063, 0.7872020263671875, 0.7871979370117187, 0.7872655639648437, 0.7871344604492188, 0.7875747680664062, 0.7872235717773437, 0.7872450561523437, 0.787177490234375, 0.7873760986328125, 0.787162109375, 0.7870628051757812, 0.7869501342773437, 0.7871918334960938, 0.7872214965820312, 0.7871826171875, 0.7872737426757812, 0.78723583984375, 0.7870596923828125, 0.7872890625, 0.7871928100585938, 0.7872849731445313, 0.7872379150390625, 0.7874027709960938, 0.7871641845703125, 0.7871498413085938, 0.7873095703125, 0.7872429809570313, 0.7872481079101562, 0.7872973022460937, 0.7877294311523437, 0.7872665405273438, 0.7872921752929688, 0.78738330078125, 0.7872399291992187, 0.78740478515625, 0.7873894653320312, 1.6892119140625, 0.7867801513671875, 0.7874058227539062, 0.787093505859375, 0.7871375122070312, 0.7874641723632813, 0.7870320434570313, 0.7867852783203125, 0.7870791625976562, 0.7871867065429687, 0.7870504760742187, 0.7871549682617187, 0.7870392456054688, 0.7870269165039062, 0.7870709838867187, 0.7869685668945312, 0.7870894165039063, 0.7870361328125, 0.78729931640625, 0.7869368286132813, 0.7870637817382813, 0.7870525512695312, 0.78691943359375, 0.7871181030273438, 0.7872645263671875, 0.787114013671875, 0.7869951782226563, 0.78718359375, 0.7872911376953124, 0.787188720703125, 0.7874611206054688, 0.7871201171875, 0.7871590576171875, 0.7871764526367188, 0.7872604370117188, 0.7872041015625, 0.787367919921875, 0.787430419921875, 0.7872379150390625, 0.7872767944335938, 0.7871437377929688, 0.7872286376953125, 0.7871508178710938, 0.7872225341796875, 0.7872214965820312, 0.78721435546875, 0.7874846801757812, 0.787336181640625, 0.787209228515625, 0.7872471313476562, 0.78731982421875, 0.787367919921875, 0.7874590454101562, 0.7876321411132813, 0.787188720703125, 0.7873546142578125, 0.7873054809570312, 0.7873382568359375, 0.7870945434570312, 0.7873085327148438, 0.7871641845703125, 0.787177490234375, 0.7871334228515625, 1.6885872802734374, 0.7868467407226563, 0.7870679321289062, 0.78702490234375, 0.7873843383789062, 0.7869224853515625, 0.7869368286132813, 0.7871651611328125, 0.7870330810546875, 0.7871365356445312, 0.7869951782226563, 0.787177490234375, 0.7871631469726562, 0.78697265625, 0.7870187377929687, 0.7872450561523437, 0.7869921264648437, 0.7871190795898437, 0.7870576782226563, 0.7870761108398437, 0.78708837890625, 0.7871754150390625, 0.7870689086914062, 0.787504150390625, 0.7870955810546875, 0.7871426391601563, 0.7869767456054687, 0.7875563354492188, 0.7872174072265625, 0.7870545654296875, 0.7871416015625, 0.787072021484375, 0.7871344604492188, 0.7871682739257813, 0.7871314086914063, 0.787078125, 0.7870269165039062, 0.7872860107421875, 0.7870443725585937, 0.7871928100585938, 0.787251220703125, 0.7872686157226563, 0.7878246459960937, 0.7873013916015625, 0.7872030639648437, 0.787198974609375, 0.7872061157226562, 0.78739453125, 0.7873382568359375, 0.7876372680664062, 0.787146728515625, 0.787293212890625, 0.7874549560546875, 0.7871702880859375, 0.7872808837890625, 0.7872481079101562, 0.7872327880859376, 0.7873157348632812, 0.7872184448242188, 0.7873003540039063, 0.7872634887695312, 0.7872921752929688, 0.7874928588867187]",tokens/s,1.2499180124325768,, 4bit-gptq-exllama-v1-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,m,m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) @@ -5416,7 +5384,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8b28-6a1657ec78b920a14d201b5f;850e189b-abc3-41ee-9120-7b1c6c19af59) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e422-6368faee2a69faa659052e39;78261a00-53e7-49a8-b092-66a4831706ec) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5486,7 +5454,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f9007-631ba437431e06dd7313028f;537f38d4-ef4d-490c-a465-35ae81906a77) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e8c6-5ed177994f68f8c37f435b66;cb751be2-b935-4c13-9733-9fc5e8993516) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5556,7 +5524,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8fb0-5afc0636751fb77f0bacb733;dbea42bc-604b-4d96-b78f-bf300bccf624) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e85c-6f55982815d9ab5656deb828;895dd110-0f05-4ef6-bd82-617cf631a08a) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5684,7 +5652,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8e91-559c538847f0117717b8e083;30b9fb39-85c0-4d93-a4d3-d1cc21b8d2be) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e744-624d3e654ad9a6230b8b6201;4d60829d-6e37-477a-b6d7-77cd884d3b8d) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5755,7 +5723,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8f14-31dd46df49cc78a72e8f38eb;364ed7f4-9b21-4d8a-b192-8b0897182dc5) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e7bd-1082501e2c7375471e486069;c606cf57-e9f0-47ec-9d39-04b4ab229bb3) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5967,7 +5935,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666f8810-6f08f32b5e9eb08b33651122;2cbca539-160c-4deb-a140-d25c07233692) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d815-0df933c6614bd43b0647cdd5;799aac47-7107-45e3-beb8-543600571bda) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -6056,7 +6024,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8ebb-04500958316fb4bc0a503765;de90e732-cfab-4e6a-aeb9-149b554e5819) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e76d-3ace38d106192f36169a22c6;2ba6e9be-4d89-4762-8af5-c929a5a87c9f) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -6116,7 +6084,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666f87dd-45978ff821160da82570b026;907f0581-df14-4ac2-a824-d8ce958a9d72) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d7ec-2ea8cc481e880d08729bd4c3;3f919044-c952-46fd-839e-69a0c748d790) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -6445,7 +6413,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666f8551-5c2a883918b01e861fea498c;d7b72526-5b6a-4d29-9a88-8f33966a4c50) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d653-278913bc5fcf8c253734c9b8;39569a41-e614-4365-ae7a-b3dbe2f536a7) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -6530,7 +6498,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8e38-4718e93768ec797040e84cca;4b67b1b9-5373-4f67-8f1d-9ab694701cb3) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e6f4-15ce0a8d2753be140373e6f7;febfa986-d5a5-4188-8368-ccc67557dc46) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -6646,7 +6614,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8d47-75f029d018eb732c77d172a5;89a59fbb-3f21-43b3-9dd9-f66a3a619ab8) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e62b-7bc0d91a28c2e4cc69be0a39;4d068fd4-3340-4559-a77c-82a38f1bc932) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -6746,7 +6714,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp61dvi4ly/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmps3ykffd7/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -6818,7 +6786,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8f99-2b0c48117c2343ed43a98b2c;d5e5c233-9f11-4d4c-b9c8-77d695846ba6) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e846-34f6aeb84ec5b559427a1d1c;5f02c05d-0996-46a9-bebb-8585927801bf) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -6875,7 +6843,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpgq5q12j0/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmprh10k01q/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,M,M,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -6917,7 +6885,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8d17-1cc54eed39d5b80e7e9b2f9c;aa646a1b-d224-4ed3-b14a-b027ac127ec7) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e602-619eba1a4b1e950c22186d0d;d05b1b81-1ab3-436b-bb83-67e5a654ec3d) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -6988,7 +6956,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8de4-2e3b7e66313660fc1d013425;a8e1a8cf-24e9-4ae3-93db-74e2d2fa899f) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e6a3-29b3da8035dda9d864f74a0e;82a1f71d-ab2a-4aa2-aba8-e042570f73b9) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7046,7 +7014,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpwt5m8ruy/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmplwwieoyi/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,4755.23072,21498.42944,0.0,20851.982336,20236.60032,s,10,15.643266601562498,1.5643266601562498,0.001716851893423332,1.563902587890625,1.5660734985351563,1.567352362060547,1.5683754528808596,"[1.5631807861328124, 1.5647735595703125, 1.5629490966796875, 1.56462890625, 1.563903564453125, 1.5627757568359375, 1.565789306640625, 1.563901611328125, 1.5686312255859376, 1.5627327880859374]",tokens/s,163.6486844598237,kWh,1.846329511867629e-05,1.0117909425098331e-05,9.183712902519626e-05,0.00012041833356897088,tokens/kWh,2125922.1284055826,MB,4755.23072,21498.42944,0.0,20851.982336,20339.706368,s,10,879.77465625,87.97746562500001,0.0022909340685085054,87.97753515625,87.97981953125,87.980624609375,87.981268671875,"[87.974140625, 87.978578125, 87.9789609375, 87.9772265625, 87.979640625, 87.97715625, 87.97784375, 87.97353125, 87.9814296875, 87.9761484375]",tokens/s,0.716092462455496,kWh,0.0010385792660216491,0.0005692333667929961,0.005151564232359207,0.006759376865173853,tokens/kWh,9320.38577765846,,s,629,892.4578986816405,1.4188519851854382,0.185906583948783,1.396368408203125,1.3966589599609376,1.3968528564453124,2.960364462890625,"[1.39618505859375, 1.39667041015625, 1.3964708251953124, 1.396378662109375, 1.3963243408203125, 1.3962271728515625, 1.3963038330078126, 1.3964288330078125, 1.3964349365234374, 1.396369384765625, 1.3964697265625, 1.3965823974609375, 1.3962998046875, 1.3964708251953124, 1.3964349365234374, 1.3963857421875, 1.396304931640625, 1.3962802734375, 1.3960714111328125, 1.396305908203125, 1.3961553955078125, 1.3963427734375, 1.396221923828125, 1.3963756103515625, 1.39614111328125, 1.396317138671875, 1.3963294677734375, 1.3963673095703124, 1.396279296875, 1.3963018798828124, 1.3959813232421876, 1.3964195556640624, 1.396304931640625, 1.396304931640625, 1.396275146484375, 1.3961236572265625, 1.3962794189453125, 1.396401123046875, 1.39871435546875, 1.396484130859375, 1.396558837890625, 1.396505615234375, 1.3966868896484375, 1.3962772216796875, 1.3962833251953124, 1.3960928955078125, 1.396380615234375, 1.3968824462890626, 1.3964442138671875, 1.396285400390625, 1.3964237060546876, 1.396358154296875, 1.396305908203125, 1.3966787109375, 1.3965537109375, 1.3964185791015624, 1.3962916259765625, 1.3968353271484375, 1.396358154296875, 1.396221923828125, 1.39633251953125, 1.396252685546875, 2.962231201171875, 1.396673583984375, 1.3970401611328125, 1.396295654296875, 1.3965557861328124, 1.396348876953125, 1.3965035400390624, 1.3961922607421875, 1.3968536376953125, 1.39629052734375, 1.396274169921875, 1.396199462890625, 1.3962281494140625, 1.3965833740234375, 1.396538330078125, 1.3962445068359375, 1.3963079833984375, 1.3963243408203125, 1.396515869140625, 1.3962762451171875, 1.396073486328125, 1.3978162841796875, 1.396114501953125, 1.396137939453125, 1.3962198486328126, 1.3962987060546874, 1.3961595458984375, 1.3963714599609376, 1.396452392578125, 1.3961502685546876, 1.3962720947265626, 1.3963868408203124, 1.3964185791015624, 1.3962239990234375, 1.396464599609375, 1.3964482421875, 1.396427734375, 1.396380615234375, 1.396588623046875, 1.3964810791015625, 1.3967718505859374, 1.3981634521484374, 1.3962802734375, 1.396275146484375, 1.3962301025390624, 1.396253662109375, 1.396327392578125, 1.3961728515625, 1.3966571044921876, 1.396474853515625, 1.3965128173828125, 1.3963858642578124, 1.396339599609375, 1.3961318359375, 1.3965147705078125, 1.3963387451171876, 1.396506591796875, 1.3962393798828125, 1.3964605712890625, 1.3965179443359375, 1.39639599609375, 1.3985382080078126, 1.3965301513671875, 2.9603779296875, 1.396199462890625, 1.3967340087890625, 1.3962291259765625, 1.3964554443359376, 1.396252685546875, 1.3964892578125, 1.3960601806640625, 1.39671240234375, 1.3963714599609376, 1.3963509521484374, 1.3961092529296875, 1.39653125, 1.3965814208984375, 1.3964400634765626, 1.3963172607421874, 1.39658349609375, 1.3963939208984375, 1.396515869140625, 1.3962587890625, 1.3965823974609375, 1.3966663818359375, 1.396378662109375, 1.3987901611328124, 1.3961759033203125, 1.3963653564453125, 1.3963018798828124, 1.3964093017578125, 1.39652197265625, 1.3961544189453126, 1.3962332763671874, 1.39641650390625, 1.39623828125, 1.3962393798828125, 1.3961964111328125, 1.396336669921875, 1.3964564208984376, 1.3965137939453125, 1.3963756103515625, 1.3962701416015626, 1.396137939453125, 1.396569091796875, 1.39635205078125, 1.3963131103515625, 1.396252685546875, 1.3963427734375, 1.3963970947265625, 1.3995396728515626, 1.39681591796875, 1.396443115234375, 1.396485107421875, 1.3962608642578125, 1.3962506103515624, 1.3963192138671876, 1.3964769287109375, 1.396210693359375, 1.3968311767578125, 1.396547607421875, 1.39646875, 1.39649951171875, 1.396304931640625, 1.3964041748046876, 1.3961563720703125, 2.961944580078125, 1.39614208984375, 1.396427734375, 1.39610009765625, 1.3963775634765625, 1.397116943359375, 1.3965689697265624, 1.396379638671875, 1.396430908203125, 1.3963284912109375, 1.3966080322265626, 1.3961544189453126, 1.3964501953125, 1.396547607421875, 1.396279296875, 1.3963203125, 1.3963448486328125, 1.396454345703125, 1.3965128173828125, 1.396253662109375, 1.396316162109375, 1.396046875, 1.396062255859375, 1.3964974365234375, 1.396443115234375, 1.3962864990234376, 1.3960980224609374, 1.3964810791015625, 1.3962884521484376, 1.396199462890625, 1.396241455078125, 1.3960233154296875, 1.3961175537109376, 1.396675537109375, 1.3962762451171875, 1.3964237060546876, 1.3964083251953125, 1.3962884521484376, 1.3965762939453126, 1.3964708251953124, 1.396631591796875, 1.3964022216796874, 1.396336669921875, 1.396317138671875, 1.400958984375, 1.396559814453125, 1.3964820556640625, 1.39627001953125, 1.3966888427734374, 1.3963427734375, 1.396474853515625, 1.3962342529296874, 1.3963448486328125, 1.39620654296875, 1.396552734375, 1.3967205810546874, 1.3966029052734374, 1.3963560791015626, 1.3962332763671874, 1.3963212890625, 1.3964134521484375, 1.396279296875, 1.39620654296875, 2.96119921875, 1.3962393798828125, 1.39626806640625, 1.3964124755859375, 1.3966131591796875, 1.396216796875, 1.3965548095703124, 1.3961031494140625, 1.396580322265625, 1.396380615234375, 1.3969398193359375, 1.396273193359375, 1.39626904296875, 1.39641650390625, 1.3962547607421876, 1.39647900390625, 1.396178955078125, 1.3985556640625, 1.3966253662109376, 1.3968516845703125, 1.3966131591796875, 1.3960867919921875, 1.39594140625, 1.3963447265625, 1.396115478515625, 1.396316162109375, 1.3962301025390624, 1.3962618408203125, 1.3963406982421875, 1.3959669189453126, 1.396400146484375, 1.3961195068359376, 1.3963663330078124, 1.396368408203125, 1.3962496337890624, 1.39649951171875, 1.3963284912109375, 1.3964493408203125, 1.3962188720703126, 1.396115478515625, 1.3963602294921875, 1.3963663330078124, 1.3966182861328125, 1.3964237060546876, 1.3963038330078126, 1.3963653564453125, 1.39626806640625, 1.396369384765625, 1.3965137939453125, 1.3962486572265624, 1.3964317626953124, 1.3962711181640626, 1.39637353515625, 1.3964647216796875, 1.39634375, 1.3963253173828125, 1.396611083984375, 1.4011822509765626, 1.3966837158203125, 1.3966038818359374, 1.39641650390625, 1.3965537109375, 1.3965322265625, 2.960329833984375, 1.3966304931640625, 1.396526123046875, 1.3960919189453125, 1.396494384765625, 1.39626904296875, 1.3971036376953125, 1.39640625, 1.3964461669921875, 1.3961175537109376, 1.3965537109375, 1.3965025634765624, 1.3961400146484375, 1.396515869140625, 1.396490234375, 1.396189208984375, 1.3961563720703125, 1.3964564208984376, 1.3965404052734376, 1.396285400390625, 1.3961553955078125, 1.396074462890625, 1.395947509765625, 1.3964676513671874, 1.3965732421875, 1.3963448486328125, 1.3962977294921874, 1.3964267578125, 1.3964461669921875, 1.39627001953125, 1.3965732421875, 1.396262939453125, 1.39629052734375, 1.39640625, 1.3961185302734376, 1.3963243408203125, 1.3963826904296874, 1.3964442138671875, 1.400627197265625, 1.396809814453125, 1.396582275390625, 1.396357177734375, 1.39643798828125, 1.3963756103515625, 1.3962301025390624, 1.3963294677734375, 1.39649951171875, 1.3963336181640624, 1.3967288818359376, 1.396634521484375, 1.3963294677734375, 1.3961072998046875, 1.3963233642578126, 1.3963182373046874, 1.39650146484375, 1.396305908203125, 1.3965128173828125, 1.396368408203125, 1.3962659912109374, 1.39628955078125, 1.3963355712890626, 1.39620654296875, 1.39628955078125, 2.96319189453125, 1.39616259765625, 1.3965670166015625, 1.39634375, 1.39635205078125, 1.3960509033203126, 1.39610107421875, 1.3959833984375, 1.3964052734375, 1.3961861572265626, 1.39652197265625, 1.39647998046875, 1.3962496337890624, 1.3968414306640624, 1.3966519775390625, 1.3960438232421875, 1.3962802734375, 1.3961964111328125, 1.3965936279296876, 1.3964493408203125, 1.3962025146484376, 1.396211669921875, 1.3961871337890626, 1.3990369873046875, 1.39656396484375, 1.39658447265625, 1.3963192138671876, 1.3963756103515625, 1.396262939453125, 1.3963499755859374, 1.3965291748046875, 1.3964564208984376, 1.396116455078125, 1.3963765869140625, 1.396496337890625, 1.396341796875, 1.3961666259765626, 1.396453369140625, 1.396306884765625, 1.3964974365234375, 1.396357177734375, 1.396156494140625, 1.396138916015625, 1.39642578125, 1.3961502685546876, 1.396420654296875, 1.3963642578125, 1.39631201171875, 1.396749267578125, 1.396473876953125, 1.396526123046875, 1.396474853515625, 1.3964871826171874, 1.396432861328125, 1.396547607421875, 1.3964410400390626, 1.4002022705078125, 1.3966510009765625, 1.39624658203125, 1.396135986328125, 1.3961492919921874, 1.3963140869140624, 1.39603759765625, 2.96123388671875, 1.3961964111328125, 1.3963387451171876, 1.3962198486328126, 1.39618408203125, 1.3961973876953124, 1.3963714599609376, 1.396177978515625, 1.3965966796875, 1.396106201171875, 1.396715576171875, 1.3962823486328124, 1.3963294677734375, 1.3963028564453126, 1.396380615234375, 1.3961964111328125, 1.398497314453125, 1.3963970947265625, 1.3965679931640624, 1.396317138671875, 1.3965782470703125, 1.3963919677734375, 1.3962659912109374, 1.3962720947265626, 1.396201416015625, 1.396317138671875, 1.3962393798828125, 1.396137939453125, 1.3964708251953124, 1.3964288330078125, 1.3963857421875, 1.3960755615234375, 1.3963480224609375, 1.396232177734375, 1.3961451416015624, 1.3963182373046874, 1.396641845703125, 1.3966080322265626, 1.396378662109375, 1.396273193359375, 1.3963345947265624, 1.3962322998046874, 1.3966744384765626, 1.3964073486328126, 1.3963243408203125, 1.3962373046875, 1.3963499755859374, 1.3963345947265624, 1.3965465087890625, 1.396432861328125, 1.39626904296875, 1.39639404296875, 1.3961748046875, 1.3964892578125, 1.3966878662109374, 1.396232177734375, 1.396611083984375, 1.3963765869140625, 1.39654248046875, 1.39650048828125, 1.39634375, 1.3964093017578125, 1.3962454833984375, 2.96591357421875, 1.3961614990234374, 1.396864013671875, 1.3961820068359374, 1.396379638671875, 1.3964666748046874, 1.3964503173828124, 1.3963642578125, 1.39664892578125, 1.39633154296875, 1.396452392578125, 1.3962650146484374, 1.396241455078125, 1.3963756103515625, 1.3963775634765625, 1.396111328125, 1.3964503173828124, 1.3980078125, 1.3967994384765625, 1.396642822265625, 1.3961103515625, 1.3965987548828125, 1.396348876953125, 1.3962291259765625, 1.396262939453125, 1.39631201171875, 1.3963212890625, 1.3961759033203125, 1.3965260009765625, 1.3965496826171875, 1.396384765625, 1.3963182373046874, 1.396358154296875, 1.3962158203125, 1.3962720947265626, 1.396420654296875, 1.3962281494140625, 1.3967728271484374, 1.3964512939453124, 1.39666845703125, 1.3965936279296876, 1.3961728515625, 1.39624853515625, 1.3961861572265626, 1.39656396484375, 1.3964697265625, 1.396485107421875, 1.396537353515625, 1.3965670166015625, 1.3964073486328126, 1.3962578125, 1.3964390869140626, 1.3966990966796875, 1.39630078125, 1.3968045654296875, 1.3967421875, 1.3966285400390626, 1.3964482421875, 1.396533203125, 1.396506591796875, 1.396284423828125, 1.3962977294921874, 1.400701904296875, 2.960291748046875, 1.3964564208984376, 1.396536376953125, 1.396294677734375, 1.396358154296875, 1.39620458984375, 1.3963621826171875, 1.3960714111328125, 1.396537353515625, 1.396304931640625, 1.3964451904296875, 1.396430908203125, 1.3963642578125, 1.3963837890625, 1.396828125, 1.3963817138671875, 1.39660498046875, 1.396589599609375, 1.39631201171875, 1.396283447265625, 1.396177978515625, 1.39618505859375, 1.396168701171875, 1.396116455078125, 1.39642578125, 1.396654052734375, 1.399869384765625, 1.3961328125, 1.3965076904296876, 1.3962025146484376, 1.3964339599609374, 1.3961707763671876, 1.3962506103515624, 1.396093994140625, 1.3964124755859375, 1.396494384765625, 1.3964676513671874, 1.396410400390625, 1.396167724609375, 1.3962772216796875, 1.396236328125, 1.3962117919921875, 1.39633251953125, 1.396305908203125, 1.396552734375, 1.3963140869140624, 1.39652197265625, 1.396462646484375, 1.3968414306640624, 1.39652099609375, 1.396305908203125, 1.396357177734375, 1.3965721435546874, 1.39648193359375, 1.396537353515625, 1.396272216796875, 1.396579345703125, 1.3963602294921875, 1.3965322265625, 1.396306884765625, 1.3964041748046876, 1.396275146484375, 1.3961829833984376]",tokens/s,0.7047951516023033,, @@ -7079,7 +7047,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666f915a-2a0842ee7eef953e0357ef42;a945b8b6-c8a0-4a07-a5e5-0781d8b1cf2b) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e9e9-204b01563abfaf961831988a;83d1bbd7-1c19-497b-8855-f4292831d35e) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -7164,7 +7132,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8e0d-0ee93908211619754094922f;b3783a59-df2f-45ff-9956-4f551006a250) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e6cb-554fdb5b18f9644a0f1a2d88;bf5994c8-713f-491f-994e-bf458f659f0e) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7226,7 +7194,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666f87b3-3e9738315c1660bc422b62b0;c2d90fee-b01e-44e6-b3cb-ab4592936ffa) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d7c4-7477b5f82140e208522b3b6c;7bb40a3d-ad51-4a14-b980-5a2a505c5c3e) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -7353,7 +7321,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8db4-04feb5a575aa322a56e827a3;fe5ef8bc-e165-47df-ab85-a98420f1a8de) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e67b-30e88d9b17996be30fa0a4fb;783340d3-1135-40d5-a600-cbfe3899dc03) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7426,7 +7394,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f883a-655211f31a3cb37d615b162e;19f83c80-74cc-4e41-a100-1ca748898c34) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670d83d-2866593f423ab21f3e2a5af1;71ca5190-ada3-4d7b-8e72-1836ae030172) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7485,7 +7453,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpbn96xo79/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpq4pybn7x/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,B,B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -7527,7 +7495,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8f42-434cd73b5c0360be4a1aea67;1be87ee2-1c66-45ce-b47e-97ad94039934) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e7e6-73f330ad76cdb97e33ae9f80;78b902c8-ec28-4237-bc48-028cd4d03d59) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7584,7 +7552,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpwt3d77y0/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpx9wjvhnv/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,TencentARC/Mistral_Pro_8B_v0.1,TencentARC/Mistral_Pro_8B_v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,5937.815552,7053.246464,0.0,6406.79936,6098.60864,s,10,4.369107391357422,0.4369107391357422,0.0007208936334109947,0.43653796386718746,0.43786113586425784,0.4381150161743164,0.4383181204223633,"[0.438368896484375, 0.43641900634765624, 0.43653814697265625, 0.43647592163085935, 0.43767733764648437, 0.43614486694335936, 0.43780471801757814, 0.43653778076171873, 0.4368502807617187, 0.4362904357910156]",tokens/s,585.9320384442743,kWh,5.155067728913349e-06,2.8244480818023514e-06,2.4793715487132108e-05,3.2773231297847804e-05,tokens/kWh,7811252.960485814,MB,5937.815552,7055.343616,0.0,6408.896512,6157.952,s,10,245.01072265624998,24.501072265624998,0.0034318484925660374,24.5004931640625,24.5048515625,24.5070751953125,24.508854101562502,"[24.50105078125, 24.49864453125, 24.50170703125, 24.49968359375, 24.496791015625, 24.509298828125, 24.50153515625, 24.504357421875, 24.499935546875, 24.49771875]",tokens/s,2.571316035355277,kWh,0.0002892534986599056,0.000158535909402625,0.0013504422735896686,0.0017982316816521993,tokens/kWh,35034.417779869254,,s,629,248.55180075073253,0.3951538962650754,0.05191103267325261,0.38883840942382814,0.3892619018554688,0.38950277099609376,0.8256093530273438,"[0.38878106689453124, 0.38896536254882813, 0.388632568359375, 0.38873394775390624, 0.38894488525390625, 0.3886663818359375, 0.3890759582519531, 0.3891517333984375, 0.3889776611328125, 0.38876263427734375, 0.38862646484375, 0.38881890869140623, 0.3890196533203125, 0.3885875244140625, 0.38878616333007815, 0.3895060424804688, 0.38884146118164065, 0.38903399658203125, 0.389644287109375, 0.38915277099609374, 0.3891998596191406, 0.3889284973144531, 0.38931558227539065, 0.3889561462402344, 0.3889131469726563, 0.38835916137695314, 0.38861004638671875, 0.38877081298828126, 0.388674560546875, 0.3886622619628906, 0.3889428405761719, 0.38882098388671876, 0.3894353942871094, 0.3886909484863281, 0.38898175048828126, 0.38895718383789063, 0.38856500244140624, 0.3886120910644531, 0.389538818359375, 0.38871450805664065, 0.38864486694335937, 0.3886510009765625, 0.3886561279296875, 0.38851071166992185, 0.3891835021972656, 0.38876364135742186, 0.3887032470703125, 0.3888701477050781, 0.38866943359375, 0.3892899780273438, 0.3888373718261719, 0.388600830078125, 0.3887615966796875, 0.3901071472167969, 0.3886766052246094, 0.3887728576660156, 0.3887442016601563, 0.3886929931640625, 0.3890124816894531, 0.3890810852050781, 0.38887628173828126, 0.388568115234375, 0.8254780883789062, 0.3889141845703125, 0.38875750732421877, 0.38877593994140625, 0.3889254455566406, 0.3885977478027344, 0.3893872680664062, 0.38873907470703123, 0.38861721801757815, 0.3886346130371094, 0.38904217529296875, 0.3888578491210937, 0.38904013061523435, 0.38923876953125, 0.38863873291015627, 0.38872164916992186, 0.38861822509765626, 0.3884800109863281, 0.3887462463378906, 0.3886069641113281, 0.3893196716308594, 0.38932070922851564, 0.3890186157226562, 0.3887646789550781, 0.38994842529296875, 0.38882098388671876, 0.3890513916015625, 0.3891650695800781, 0.38888037109375, 0.3886991882324219, 0.3889468688964844, 0.3885434875488281, 0.3888199768066406, 0.3885844421386719, 0.388600830078125, 0.3886417846679687, 0.3889479675292969, 0.3886417846679687, 0.3886336059570312, 0.3887790222167969, 0.38892236328125, 0.38850970458984374, 0.38881793212890625, 0.3887523803710938, 0.3886868591308594, 0.38856500244140624, 0.3889899597167969, 0.38897457885742187, 0.38872164916992186, 0.3891282043457031, 0.38919577026367186, 0.3890595703125, 0.3892817993164063, 0.38935140991210937, 0.3886397399902344, 0.38855474853515626, 0.3887595520019531, 0.3887523803710938, 0.3889899597167969, 0.38894900512695313, 0.38868582153320314, 0.38889370727539063, 0.3889704895019531, 0.8256952514648438, 0.38874215698242187, 0.38912716674804687, 0.3890257873535156, 0.38883328247070315, 0.3886970825195312, 0.3889541015625, 0.38887527465820315, 0.389212158203125, 0.389137451171875, 0.38895819091796874, 0.3886970825195312, 0.3888670654296875, 0.38879641723632813, 0.38865716552734375, 0.3888598937988281, 0.3887564697265625, 0.3886991271972656, 0.38894900512695313, 0.3889837951660156, 0.38897357177734376, 0.3888793640136719, 0.38878106689453124, 0.38885479736328127, 0.3889776611328125, 0.38883840942382814, 0.38856396484375, 0.3887400817871094, 0.38860800170898435, 0.3888199768066406, 0.38868582153320314, 0.3887298583984375, 0.38889166259765623, 0.38933505249023437, 0.3892725830078125, 0.3889756164550781, 0.38903604125976565, 0.38886810302734376, 0.38863665771484374, 0.3886458740234375, 0.3889704895019531, 0.38879129028320314, 0.3888148498535156, 0.3887943725585937, 0.3887175598144531, 0.38925106811523436, 0.3889131469726563, 0.38866534423828125, 0.3887677307128906, 0.38883840942382814, 0.388853759765625, 0.388853759765625, 0.3888025512695312, 0.38896536254882813, 0.38888958740234375, 0.3887513732910156, 0.38878823852539063, 0.38919064331054687, 0.38868582153320314, 0.39011123657226565, 0.3893196716308594, 0.38929306030273436, 0.38961663818359377, 0.8260413208007813, 0.38885992431640626, 0.3889643249511719, 0.3887523803710938, 0.3887247314453125, 0.3887185974121094, 0.38895718383789063, 0.38912408447265623, 0.3887984619140625, 0.3889039306640625, 0.388822021484375, 0.3887298583984375, 0.388811767578125, 0.3887718505859375, 0.38874215698242187, 0.3888404541015625, 0.38900531005859373, 0.38877694702148435, 0.3888035888671875, 0.3886673889160156, 0.38864999389648436, 0.38919064331054687, 0.38883941650390624, 0.38895718383789063, 0.38876263427734375, 0.38873394775390624, 0.3887442016601563, 0.3888988037109375, 0.38905035400390625, 0.38884658813476564, 0.38873294067382813, 0.3889889221191406, 0.3889121398925781, 0.3887595520019531, 0.3887544250488281, 0.38894488525390625, 0.38860800170898435, 0.38863873291015627, 0.3886981201171875, 0.38873294067382813, 0.3888630065917969, 0.3888568115234375, 0.3888783264160156, 0.38862539672851565, 0.3889613037109375, 0.389233642578125, 0.38876263427734375, 0.38913946533203125, 0.3891650695800781, 0.3887237243652344, 0.38868582153320314, 0.38899713134765623, 0.3889704895019531, 0.3898193969726563, 0.3890257873535156, 0.3886673889160156, 0.38894692993164065, 0.3890390930175781, 0.3892991943359375, 0.38878515625, 0.3889059753417969, 0.3887462463378906, 0.3888424987792969, 0.826693603515625, 0.3888025512695312, 0.3887523803710938, 0.3885926513671875, 0.3887923278808594, 0.3886120910644531, 0.38872576904296874, 0.38866021728515626, 0.3889807434082031, 0.38920294189453125, 0.38865304565429687, 0.38867559814453123, 0.3889039306640625, 0.388706298828125, 0.3886264343261719, 0.3890257873535156, 0.3886796875, 0.38881893920898436, 0.388748291015625, 0.3887544250488281, 0.3898101806640625, 0.38882098388671876, 0.3889725341796875, 0.38886605834960936, 0.3887820739746094, 0.38895001220703124, 0.3888015441894531, 0.38883840942382814, 0.3887093811035156, 0.38877593994140625, 0.3885823974609375, 0.3885844421386719, 0.3888087158203125, 0.388632568359375, 0.388895751953125, 0.3889407958984375, 0.3887400817871094, 0.38869403076171877, 0.38845440673828124, 0.388706298828125, 0.3888496704101562, 0.38918142700195313, 0.3888035888671875, 0.38866534423828125, 0.3889039306640625, 0.38894488525390625, 0.3888486328125, 0.38889984130859373, 0.3886970825195312, 0.3889612731933594, 0.3889326171875, 0.38907904052734377, 0.38912612915039063, 0.388790283203125, 0.3889459228515625, 0.38889675903320314, 0.3888025512695312, 0.38870016479492187, 0.38883428955078125, 0.3888517150878906, 0.38923776245117186, 0.38900222778320315, 0.38878616333007815, 0.8254668579101563, 0.389001220703125, 0.3892408447265625, 0.38915377807617185, 0.38880767822265627, 0.38871142578125, 0.38886297607421877, 0.3891148681640625, 0.38876364135742186, 0.38882720947265625, 0.3886867980957031, 0.3885404052734375, 0.38891827392578127, 0.38905035400390625, 0.38921829223632814, 0.3889141845703125, 0.3891558532714844, 0.38889471435546874, 0.3890810852050781, 0.38909234619140626, 0.3887319030761719, 0.38957876586914064, 0.3891712036132812, 0.3887923583984375, 0.3895316162109375, 0.38924493408203126, 0.38886605834960936, 0.38888653564453124, 0.38975283813476563, 0.38893875122070315, 0.38879641723632813, 0.38891009521484377, 0.38910565185546875, 0.3892593078613281, 0.39087921142578125, 0.38871551513671876, 0.38870834350585937, 0.38860186767578125, 0.3886960754394531, 0.38911077880859374, 0.38857318115234374, 0.3887657470703125, 0.3889714660644531, 0.38878106689453124, 0.3889807434082031, 0.3888148498535156, 0.3887872009277344, 0.3890739135742188, 0.38941082763671875, 0.3891589050292969, 0.38924798583984377, 0.3888977966308594, 0.38902792358398436, 0.3890799560546875, 0.3894364013671875, 0.3890677795410156, 0.388959228515625, 0.38916094970703125, 0.3891568603515625, 0.3888988037109375, 0.38902987670898437, 0.38889370727539063, 0.38949786376953127, 0.8270551147460937, 0.38860186767578125, 0.38860595703125, 0.38868377685546873, 0.38870834350585937, 0.38866021728515626, 0.3888844909667969, 0.38888858032226564, 0.3890831298828125, 0.388947998046875, 0.38878512573242185, 0.389180419921875, 0.38898175048828126, 0.38901556396484377, 0.38881381225585937, 0.38938214111328123, 0.38876568603515627, 0.38882098388671876, 0.3885455322265625, 0.3887093811035156, 0.3890534362792969, 0.38878515625, 0.3887400817871094, 0.3889725341796875, 0.3888455810546875, 0.38887628173828126, 0.3890677795410156, 0.38898175048828126, 0.38893771362304685, 0.3888015441894531, 0.38900735473632814, 0.3888015441894531, 0.38892440795898436, 0.3890124816894531, 0.3895828552246094, 0.3893197021484375, 0.3891650390625, 0.3890083923339844, 0.38891827392578127, 0.38878311157226564, 0.3886766052246094, 0.38947021484375, 0.3891251220703125, 0.38870834350585937, 0.3887933349609375, 0.3888517150878906, 0.38911181640625, 0.38871142578125, 0.38868582153320314, 0.3887319030761719, 0.388611083984375, 0.38874215698242187, 0.388890625, 0.3886889038085938, 0.38894692993164065, 0.38870220947265627, 0.3885906066894531, 0.38906878662109373, 0.3886192626953125, 0.38860186767578125, 0.3894384765625, 0.38954290771484373, 0.3893800964355469, 0.825660400390625, 0.38914047241210936, 0.3887677307128906, 0.38940365600585936, 0.38903704833984376, 0.3888568420410156, 0.38899917602539064, 0.3896258544921875, 0.3887247314453125, 0.3890565185546875, 0.3885342712402344, 0.38850048828125, 0.38885888671875, 0.38866021728515626, 0.38881793212890625, 0.3884103698730469, 0.38903604125976565, 0.38858648681640623, 0.3889070129394531, 0.3887185974121094, 0.38888858032226564, 0.3892193298339844, 0.3894497375488281, 0.38906060791015623, 0.3886120910644531, 0.388600830078125, 0.38877490234375, 0.3888005065917969, 0.38874725341796873, 0.38922955322265623, 0.3888926696777344, 0.3885957336425781, 0.38874212646484374, 0.3889346618652344, 0.3888650207519531, 0.38876876831054685, 0.388664306640625, 0.38930841064453126, 0.3888230285644531, 0.38908517456054686, 0.38917633056640627, 0.38882611083984375, 0.3889510498046875, 0.38876058959960935, 0.3885772705078125, 0.3890186157226562, 0.3886346130371094, 0.3894353942871094, 0.389170166015625, 0.3888036193847656, 0.38960125732421874, 0.3888005065917969, 0.3895419006347656, 0.3890739135742188, 0.3897927551269531, 0.38913229370117186, 0.38885888671875, 0.3889837951660156, 0.388969482421875, 0.38935653686523436, 0.38904013061523435, 0.3888988037109375, 0.38913433837890626, 0.8261171264648437, 0.38934527587890627, 0.3890943908691406, 0.3892561950683594, 0.38904730224609374, 0.3889776611328125, 0.38878823852539063, 0.38880459594726563, 0.38874215698242187, 0.38939239501953127, 0.3890657348632813, 0.3885916137695313, 0.38864999389648436, 0.3887585144042969, 0.388790283203125, 0.3887032470703125, 0.3896012878417969, 0.3887442016601563, 0.3886336059570312, 0.38861004638671875, 0.389037109375, 0.3888260498046875, 0.388701171875, 0.38865304565429687, 0.38868069458007815, 0.38857318115234374, 0.3886213073730469, 0.3887431640625, 0.38874725341796873, 0.38863873291015627, 0.388595703125, 0.38875341796875, 0.38857522583007814, 0.3887923278808594, 0.38920806884765624, 0.3891712036132812, 0.38883123779296874, 0.3889407958984375, 0.388748291015625, 0.3887503356933594, 0.38887527465820315, 0.38860287475585936, 0.38913742065429685, 0.3887790222167969, 0.3885629577636719, 0.38873394775390624, 0.38880563354492187, 0.38863565063476563, 0.38857318115234374, 0.3885987854003906, 0.38900531005859373, 0.3888097229003906, 0.38852197265625, 0.3885834350585938, 0.38872576904296874, 0.38862234497070314, 0.3915376586914063, 0.38877694702148435, 0.3889428405761719, 0.38899917602539064, 0.38912716674804687, 0.38925726318359377, 0.3891875610351562, 0.8258375854492187, 0.388811767578125, 0.38908212280273435, 0.3887953796386719, 0.38926849365234373, 0.38856192016601565, 0.38865304565429687, 0.38886297607421877, 0.38894692993164065, 0.38873086547851565, 0.3887933349609375, 0.3884656677246094, 0.38854656982421876, 0.38860287475585936, 0.3886929931640625, 0.3886868591308594, 0.38878411865234375, 0.3888066711425781, 0.3886141357421875, 0.3886141357421875, 0.38892340087890626, 0.38862335205078125, 0.3893319702148437, 0.3891773376464844, 0.38893875122070315, 0.38863565063476563, 0.3889203186035156, 0.3887298583984375, 0.38881689453125, 0.3891302490234375, 0.3888824462890625, 0.3886489562988281, 0.3886243896484375, 0.3889407958984375, 0.38874111938476563, 0.38874111938476563, 0.38891009521484377, 0.3886141357421875, 0.38878823852539063, 0.38871142578125, 0.3888291931152344, 0.38855270385742186, 0.38883328247070315, 0.388885498046875, 0.38914764404296875, 0.38934017944335936, 0.38926025390625, 0.38955938720703126, 0.3897701416015625, 0.3890442199707031, 0.38865716552734375, 0.3886612548828125, 0.3888455810546875, 0.38867559814453123, 0.38898483276367185, 0.3890862121582031, 0.38879129028320314, 0.38848614501953127, 0.3886970825195312, 0.3888097229003906, 0.3888128356933594, 0.3887206115722656, 0.38876980590820315]",tokens/s,2.530659597315939,, @@ -7627,7 +7595,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8b9d-34617c7d6058c0d023cc7b40;13d00bdf-9af6-49b1-b436-dd00b65a14c4) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e494-46292da06990fbe7490ac973;ee31bf9d-5f1f-4982-9e82-7ee98c875a61) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7685,7 +7653,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp3i5r6q7l/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmphgmks7nm/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,facebook/opt-350m,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1269.682176,1019.74016,0.0,373.293056,277.410816,s,10,0.37923129653930665,0.037923129653930666,0.0012152886645715089,0.038048974990844725,0.038519168853759764,0.0396308479309082,0.04052019119262695,"[0.04074252700805664, 0.03821686553955078, 0.03797974395751953, 0.03823471832275391, 0.03626841735839844, 0.03608185577392578, 0.03733708953857422, 0.03811382293701172, 0.03827212905883789, 0.03798412704467773]",tokens/s,6750.497713035296,kWh,4.301884120160883e-07,2.3572264180256752e-07,1.164744366137984e-06,1.83065541995664e-06,tokens/kWh,139840626.0453229,MB,1269.682176,1019.74016,0.0,373.293056,323.047424,s,10,23.493592773437502,2.3493592773437504,0.014632907695307939,2.3561588134765623,2.362567529296875,2.3635712402343754,2.3643742089843753,"[2.34352734375, 2.36090478515625, 2.362344482421875, 2.364574951171875, 2.338309326171875, 2.331607421875, 2.357506103515625, 2.3608046875, 2.3548115234375, 2.3192021484375]",tokens/s,26.81582191687153,kWh,2.642717914328431e-05,1.4482863030778391e-05,6.568572578086276e-05,0.00010659576795492543,tokens/kWh,591017.8350292469,,s,629,23.794291763305676,0.037828762739754634,0.004511848848474209,0.03739033508300781,0.03767849044799805,0.03791810684204102,0.07427948760986337,"[0.03534745788574219, 0.03654348754882813, 0.03639910507202149, 0.03678003311157227, 0.03678003311157227, 0.036736000061035154, 0.036125694274902344, 0.03654348754882813, 0.036519935607910156, 0.0366110725402832, 0.03650764846801758, 0.03670425415039062, 0.03646771240234375, 0.037282817840576174, 0.03782860946655273, 0.03862527847290039, 0.037749759674072264, 0.03757056045532227, 0.037465087890625, 0.0372408332824707, 0.03734630584716797, 0.03750092697143555, 0.03727872085571289, 0.03741388702392578, 0.03740262222290039, 0.03726131057739258, 0.037303295135498044, 0.03731353759765625, 0.037188640594482424, 0.037268447875976565, 0.03742617416381836, 0.03730124664306641, 0.03739136123657227, 0.037250049591064455, 0.03735039901733399, 0.03733913421630859, 0.03732787322998047, 0.03716505432128906, 0.03726233673095703, 0.0372490234375, 0.03732070541381836, 0.03729305648803711, 0.03743129730224609, 0.03723980712890625, 0.037269504547119144, 0.03735039901733399, 0.03727769470214844, 0.03724800109863281, 0.03734124755859375, 0.037339073181152344, 0.037269504547119144, 0.03736678314208984, 0.0371701774597168, 0.037266433715820314, 0.03737190246582031, 0.037318656921386716, 0.03726847839355469, 0.03745894241333008, 0.03732787322998047, 0.03729715347290039, 0.037501953125, 0.037501953125, 0.07589580535888672, 0.03728793716430664, 0.037407745361328126, 0.037207038879394534, 0.03731353759765625, 0.03784089660644531, 0.037353473663330077, 0.03739340972900391, 0.0374015998840332, 0.03751731109619141, 0.03856486511230469, 0.037754878997802735, 0.037369857788085936, 0.03734220886230469, 0.03731353759765625, 0.0373125114440918, 0.03733606338500976, 0.037425151824951174, 0.037303295135498044, 0.037362686157226564, 0.038662143707275394, 0.037748737335205076, 0.03751116943359375, 0.03752243041992188, 0.03749785614013672, 0.03739340972900391, 0.03728793716430664, 0.03726540756225586, 0.03750707244873047, 0.03739340972900391, 0.03738828659057617, 0.03735859298706055, 0.037395454406738284, 0.03731558227539063, 0.037868545532226565, 0.037456897735595705, 0.03736883163452148, 0.03723161697387695, 0.03758489608764649, 0.03761151885986328, 0.03742822265625, 0.037384193420410154, 0.03741900634765625, 0.03744976043701172, 0.03728688049316406, 0.03746815872192383, 0.03732275390625, 0.037525505065917966, 0.0375203857421875, 0.03739340972900391, 0.03733913421630859, 0.03781631851196289, 0.03749785614013672, 0.03746815872192383, 0.03744870376586914, 0.037324798583984374, 0.03752345657348633, 0.03728691101074219, 0.0374015998840332, 0.03738828659057617, 0.037454849243164064, 0.037454849243164064, 0.03749273681640625, 0.0757176284790039, 0.03736166381835938, 0.03737702560424805, 0.03760332870483398, 0.03744768142700195, 0.037354496002197264, 0.03724595260620117, 0.037343231201171875, 0.03800678253173828, 0.03777740859985351, 0.03753472137451172, 0.03747430419921875, 0.0374835205078125, 0.037585918426513674, 0.037338111877441404, 0.0374466552734375, 0.03750912094116211, 0.037379070281982424, 0.03765964889526367, 0.03735551834106445, 0.03742310333251953, 0.03738726425170898, 0.03740364837646484, 0.037495807647705076, 0.03738214492797851, 0.037310462951660156, 0.03733401489257813, 0.03747430419921875, 0.03743436813354492, 0.0373831672668457, 0.03754393768310547, 0.037370880126953124, 0.03760025787353516, 0.037601280212402347, 0.037510143280029294, 0.0373043212890625, 0.0373125114440918, 0.03747225570678711, 0.03762688064575195, 0.037282817840576174, 0.037294078826904296, 0.03753267288208008, 0.037607425689697264, 0.037400577545166014, 0.037495807647705076, 0.03745075225830078, 0.038043647766113284, 0.037773311614990236, 0.03741798400878906, 0.03741183853149414, 0.03771289443969727, 0.03767193603515625, 0.03745280075073242, 0.0372674560546875, 0.03750092697143555, 0.038141952514648435, 0.037969921112060545, 0.037384193420410154, 0.03739852905273437, 0.03727977752685547, 0.03753366470336914, 0.037572608947753904, 0.03738726425170898, 0.07624192047119141, 0.03899084854125977, 0.037817344665527344, 0.03795356750488281, 0.03747836685180664, 0.03777536010742188, 0.03754905700683594, 0.03737497711181641, 0.03731763076782227, 0.03754598236083984, 0.0381317138671875, 0.037395454406738284, 0.03748556900024414, 0.03738726425170898, 0.037493759155273435, 0.037318656921386716, 0.03763916778564453, 0.03733299255371094, 0.037407745361328126, 0.03729817581176758, 0.03746918487548828, 0.03743129730224609, 0.03740364837646484, 0.03732275390625, 0.03753267288208008, 0.03765350341796875, 0.03748863983154297, 0.03742617416381836, 0.03767708969116211, 0.03730940628051758, 0.03756748962402344, 0.03845017623901367, 0.03810201644897461, 0.03751321411132812, 0.03726847839355469, 0.03728998565673828, 0.03749273681640625, 0.03731148910522461, 0.03733606338500976, 0.03725107192993164, 0.03753881454467774, 0.03738726425170898, 0.037282817840576174, 0.03728179168701172, 0.0374917106628418, 0.037351425170898435, 0.03746819305419922, 0.037434337615966796, 0.037440513610839846, 0.03745587158203125, 0.03747532653808594, 0.03769651031494141, 0.03751935958862305, 0.037397502899169925, 0.03741491317749023, 0.037466110229492186, 0.03749478530883789, 0.037566463470458986, 0.03733401489257813, 0.03742822265625, 0.03756851196289063, 0.037501953125, 0.03759820938110352, 0.07184793853759766, 0.03510374450683594, 0.03521535873413086, 0.03516928100585937, 0.03511603164672852, 0.03513241577148438, 0.03520000076293945, 0.035148799896240236, 0.03514470291137695, 0.03511500930786133, 0.0367718391418457, 0.03750400161743164, 0.03728998565673828, 0.037482494354248046, 0.0374835205078125, 0.03756032180786133, 0.037303295135498044, 0.037424129486083986, 0.03768729782104492, 0.03739648056030274, 0.03758492660522461, 0.03751830291748047, 0.03748659133911133, 0.03748863983154297, 0.037410816192626956, 0.03736883163452148, 0.03754396820068359, 0.03754390335083008, 0.037591041564941405, 0.03752140808105469, 0.03739136123657227, 0.03755212783813477, 0.03739340972900391, 0.03741491317749023, 0.03769139099121094, 0.037408767700195314, 0.038091777801513675, 0.03745382308959961, 0.0376360969543457, 0.03739136123657227, 0.03732070541381836, 0.03746303939819336, 0.03743334579467773, 0.037364734649658206, 0.037526527404785154, 0.0374015998840332, 0.037416961669921874, 0.037416961669921874, 0.03750604629516602, 0.037427200317382815, 0.037367809295654295, 0.037498878479003905, 0.037397502899169925, 0.03749683380126953, 0.037622814178466794, 0.03773948669433594, 0.037416961669921874, 0.03744464111328125, 0.03751728057861328, 0.0374835205078125, 0.03757056045532227, 0.037477375030517575, 0.037720062255859374, 0.0716236801147461, 0.035062782287597655, 0.03518668746948242, 0.035019775390625, 0.03531980895996094, 0.035253246307373046, 0.03521843338012695, 0.034840576171875, 0.03494911956787109, 0.03526863861083984, 0.03520406341552734, 0.035253246307373046, 0.03508428955078125, 0.0372490234375, 0.037318656921386716, 0.03726335906982422, 0.03732787322998047, 0.03786342239379883, 0.037288959503173826, 0.0374835205078125, 0.03732992172241211, 0.03781631851196289, 0.038765567779541016, 0.03768012619018555, 0.03770982360839844, 0.037367809295654295, 0.03745894241333008, 0.03721318435668945, 0.03721420669555664, 0.03736678314208984, 0.03769343948364258, 0.03735756683349609, 0.03937484741210937, 0.03793203353881836, 0.037556224822998044, 0.03736883163452148, 0.03729510498046875, 0.03745177459716797, 0.03745792007446289, 0.037561344146728515, 0.037422080993652344, 0.037353473663330077, 0.037594112396240234, 0.03732070541381836, 0.03734732818603516, 0.037340160369873046, 0.037397502899169925, 0.037362686157226564, 0.03757875061035156, 0.03736064147949219, 0.037321727752685545, 0.03742822265625, 0.03740364837646484, 0.037318656921386716, 0.03724492645263672, 0.03730022430419922, 0.03753472137451172, 0.0374200325012207, 0.03728486251831055, 0.03722751998901367, 0.03739136123657227, 0.037362686157226564, 0.037375999450683595, 0.07522509002685547, 0.03725209426879883, 0.03725823974609375, 0.0374466552734375, 0.03733916854858398, 0.03731145477294922, 0.03730124664306641, 0.03746406555175781, 0.037375999450683595, 0.03750707244873047, 0.037292030334472655, 0.037459968566894535, 0.03742924880981445, 0.037405696868896485, 0.03738009643554688, 0.0372592658996582, 0.03754905700683594, 0.03736067199707031, 0.03739644622802735, 0.037394432067871096, 0.037394432067871096, 0.037392383575439454, 0.037389312744140625, 0.0372674560546875, 0.03746713638305664, 0.037416961669921874, 0.03733606338500976, 0.03751321411132812, 0.03757056045532227, 0.037351425170898435, 0.03736883163452148, 0.037335041046142575, 0.03735756683349609, 0.03738521575927734, 0.03734630584716797, 0.03721932983398438, 0.037397502899169925, 0.03740364837646484, 0.037465087890625, 0.03727974319458008, 0.03751116943359375, 0.03736883163452148, 0.03720294570922852, 0.03728179168701172, 0.03764019012451172, 0.037395454406738284, 0.037294078826904296, 0.037379070281982424, 0.03743231964111328, 0.03737705612182617, 0.03722953414916992, 0.03727360153198242, 0.03775692749023438, 0.03735039901733399, 0.03786652755737305, 0.03738211059570312, 0.03764019012451172, 0.03744255828857422, 0.03744153594970703, 0.03740671920776367, 0.038042625427246096, 0.03756032180786133, 0.03753267288208008, 0.07615692901611328, 0.03728486251831055, 0.03751628875732422, 0.03750092697143555, 0.03723161697387695, 0.03742105484008789, 0.03728384017944336, 0.03728793716430664, 0.03723369598388672, 0.037351390838623044, 0.0372592658996582, 0.03703603363037109, 0.03736576080322265, 0.03826483154296875, 0.03825356674194336, 0.037498878479003905, 0.03745075225830078, 0.03750604629516602, 0.03733606338500976, 0.0373493766784668, 0.037405696868896485, 0.037548030853271484, 0.037897216796875, 0.0374015998840332, 0.03749273681640625, 0.03787980651855469, 0.03769548797607422, 0.03733401489257813, 0.037424129486083986, 0.03724288177490234, 0.03732275390625, 0.03731353759765625, 0.037348350524902346, 0.0373043212890625, 0.03746099090576172, 0.03751731109619141, 0.037424129486083986, 0.037477375030517575, 0.03738623809814453, 0.037395454406738284, 0.0372828483581543, 0.0373001937866211, 0.03731455993652344, 0.03749683380126953, 0.037678081512451174, 0.03893964767456055, 0.037498878479003905, 0.03726847839355469, 0.03724595260620117, 0.037381118774414065, 0.03728998565673828, 0.03745177459716797, 0.03747635269165039, 0.037425151824951174, 0.03755724716186523, 0.03739033508300781, 0.03731763076782227, 0.03747532653808594, 0.03732070541381836, 0.03763302230834961, 0.0374917106628418, 0.03746303939819336, 0.037392383575439454, 0.07604838562011719, 0.03738726425170898, 0.03732275390625, 0.037763072967529294, 0.03760537719726562, 0.037493759155273435, 0.03744153594970703, 0.037367809295654295, 0.037353473663330077, 0.03748966217041016, 0.037321727752685545, 0.03739136123657227, 0.03747020721435547, 0.03746201705932617, 0.03758899307250976, 0.037288959503173826, 0.03742617416381836, 0.037359615325927735, 0.03741900634765625, 0.03733708953857422, 0.0372592658996582, 0.03737497711181641, 0.03758182525634766, 0.03737395095825195, 0.03727052688598633, 0.03736678314208984, 0.037477375030517575, 0.03746201705932617, 0.03737702560424805, 0.03690291213989258, 0.03730022430419922, 0.037541889190673826, 0.03752140808105469, 0.03749683380126953, 0.03740979385375977, 0.037354496002197264, 0.037292030334472655, 0.03724288177490234, 0.03722956848144531, 0.03722956848144531, 0.0375203857421875, 0.03734630584716797, 0.03717836761474609, 0.0371671028137207, 0.037372928619384765, 0.037341182708740234, 0.03730636978149414, 0.03713945770263672, 0.03719987106323242, 0.03745177459716797, 0.03732889556884766, 0.03736166381835938, 0.0373493766784668, 0.03739648056030274, 0.03730739212036133, 0.03730739212036133, 0.03730944061279297, 0.03724390411376953, 0.037353473663330077, 0.03740467071533203, 0.0375203857421875, 0.03734630584716797, 0.03746406555175781, 0.07565721893310547, 0.0372674560546875, 0.0373125114440918, 0.037479423522949216, 0.03725721740722656, 0.03723468780517578, 0.0374200325012207, 0.03733196640014649, 0.03729817581176758, 0.03751321411132812, 0.03730739212036133, 0.03743334579467773, 0.03721932983398438, 0.03751833724975586, 0.03735244750976562, 0.037408767700195314, 0.03565260696411133, 0.035312641143798826, 0.03527782440185547, 0.0352542724609375, 0.03772415924072266, 0.0377077751159668, 0.03799244689941406, 0.03948953628540039, 0.037106689453125, 0.03671449661254883, 0.03655372619628906, 0.03666534423828125, 0.03620249557495117, 0.03514470291137695, 0.03519692611694336, 0.035335166931152344, 0.03523788833618164, 0.03600076675415039, 0.03778662490844727, 0.03708927917480469, 0.03679846572875976, 0.036724769592285156, 0.03671651077270508, 0.0366192626953125, 0.036624416351318356, 0.03651375961303711, 0.03685990524291992, 0.03674524688720703, 0.03836924743652344, 0.03679436874389649, 0.03653734588623047, 0.03676671981811523, 0.03650559997558594, 0.03653529739379883, 0.036706302642822264, 0.03682099151611328, 0.036528129577636716, 0.03685887908935547, 0.03667865753173828, 0.036674560546875, 0.036618240356445314, 0.03663052749633789, 0.03661721420288086, 0.03660595321655274, 0.0365926399230957, 0.036544513702392575, 0.036792320251464845]",tokens/s,26.434911627418618,, @@ -7849,7 +7817,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8b3d-5d4c45a65280a1944b188d3c;bf39ece5-6fd3-4d0f-91bb-bea8a9caf13a) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e438-246e3a412863bd7a132be354;6a3c9916-c47c-4ea3-ba3e-24ebb097403f) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7919,7 +7887,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f901b-1b92c0621e4a7db369f8659c;9b8f915d-bc2f-405f-a5ab-af3a6c46eb6c) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e8de-4557534862895c7557a696dc;aa30eafc-1c6d-4e6a-8963-9f656a28d49d) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7989,7 +7957,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8fc4-062c2ba8276bda1b54a000e5;b71ab8b7-4661-4d02-96a8-58a678711250) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e879-78710c6c66c07dea5cdebb7b;98ac9079-ee4d-42f5-9fef-d049796cdf26) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -8089,7 +8057,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8e6d-1d4c39655a80714707a63590;1aa7ebdd-faca-41be-a76e-262c2cbe42af) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e723-4616e4c66e7d031e35295179;aa9e5175-d162-44eb-aef6-b44e9f056b89) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -8160,7 +8128,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8eed-117090805f3f526419644ae6;db4122fe-ee77-49e1-87f3-c0874ccd7e45) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e79c-28a2aaf333ecf70349bb5bcf;50fc0d7e-7011-4a6b-96d1-6a0f82032036) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -8344,7 +8312,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666f87ec-3650f3be5878e42b028dfc21;4272608d-c1d6-4393-a8f7-96a9a3e57514) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d7f3-418d75973fe823d314735444;9833771b-a751-46fe-bb87-8c895259c2ad) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -8433,7 +8401,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8e99-021784481bb3815704149856;845ad8dd-5077-47f4-a92f-31944190739e) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e74b-577737943636402d0aa5eff9;cab60354-6b3b-4502-b086-923eff856bb7) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -8493,7 +8461,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666f87bb-7f48b56a055a03dd3d38efa3;514702ae-7305-4687-9e56-48e3d737323c) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d7cb-6424615322109faf177d642b;64895d45-121c-4009-b15c-675632db00d5) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -8879,7 +8847,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8e14-53eb93311d6481c07ca18fc7;bc08c253-b00d-40d7-95db-b94984979a85) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e6d3-40d238e37a7b956a3658228e;07afc1f0-dffa-43e0-b24a-7aa3509b092f) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -8995,7 +8963,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8d1f-32f305971ecca10179a5ab0b;feb29633-24a3-42d7-b67f-5976bfeed507) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e609-0ea35221113d393025af4796;4b04f458-d2b2-4176-8f88-4873069bfe3d) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9139,7 +9107,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8f77-73b04b922dbde40e7c7fe41f;d4f2666e-2ce5-4f24-88a3-fe5e2da10b14) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e815-3bdcd0c640944d8d6abbf70f;0f85c3a8-28d4-4a36-b2b5-b118d3dd80d4) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9210,7 +9178,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8cec-3c11e50e35c88988726eaa0e;4f753b09-179f-4af0-a214-40ed0166750e) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e5e1-3399b02f4c3630bc3377e2f4;f8eabdcd-596e-4277-85c1-1c20a02b4721) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9281,7 +9249,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8dbb-463b195017bafd9d6d184eb1;2bf38984-bfb5-42de-b521-d02a061cbe43) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e682-48b9bd1d0dc23d907558cbe4;4c318d9a-fdd6-4bcf-858e-984244999eda) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9400,7 +9368,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666f9130-2f05dc2d1f9718610ac56542;18249129-d209-42ad-a591-eaf502d8578b) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e9c8-5efca3534b5c84ab296728ad;8f35f1a7-f417-4481-9dd1-b21145250aa4) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -9485,7 +9453,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8dea-322ac474491977f26ee24ba7;ad7890d0-93a5-46e6-bd61-d9730a59525e) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e6aa-6254ea835cf5b9f1073e6690;4ebde97c-e072-4899-82b0-395c78b4112e) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9547,7 +9515,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666f8785-7c9678ee522d26e31774962c;a42dace1-bf2d-42b6-8c98-424f33bfab26) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d798-1dfc43e7409aaa5a399dea5e;89fbd57e-a826-4364-99f5-5e7c0f08ed3d) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -9674,7 +9642,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8d8b-7c08175c6398f46b3d81a92e;fd2c0fd1-6977-4601-a497-34e483d01760) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e65a-6a89801b7f1346bd64d50933;38ef97ee-5807-452e-9396-c3db0083f808) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9747,7 +9715,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8816-4b3bd181045798405a36a9ff;f7d5a8f5-60bd-4c72-85c9-e7843aa462ab) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670d81c-7cb013a2595a11281298964d;cfa49739-64a6-4d44-ab79-71904a526655) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9820,7 +9788,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8f1e-4200b3241378cbd3480438a9;1a261f36-7ec4-4e72-b761-48bac9ca1755) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e7c4-30e3aba7224c52001243cdf4;20ec4d2e-4a5f-4de6-8ab1-363eb78d4a8d) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9948,7 +9916,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8b6e-42746bbe7f16a47019aaf7a3;22a7f571-5872-48e4-a89a-f60346d75f63) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e46e-5f95bfb3137bfa91380a856c;756caa0c-0113-45c7-abef-187f3343ac26) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10142,7 +10110,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8b19-5373f2846e75b66c3b90fbc3;39c65910-aa07-4013-97b9-da49dd20f3d8) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e414-564547120400a78525181c7a;fb1c0bb6-921e-44ec-ac76-8191cbc005da) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10212,7 +10180,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8ff9-6615909b5d2bd598660631bb;409fd4c8-b8af-4ed3-8377-f6cb7ea6db07) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e8b7-79cb110405bd01690909e0b1;fb881c8d-9c62-4328-9e70-4ddc9f9f6f6f) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10282,7 +10250,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8fa0-0606e21935159644017017bb;5aae769d-e4b0-416a-8f16-a58a885c763e) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e84d-395f8e807424f414565841db;8bb28ac2-ae64-4459-b91b-562c754a1b6c) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10382,7 +10350,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8e75-51eb093143e206155fc5d09a;d83e88fa-621f-4ebe-b3ea-33bed109e263) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e72a-1c994828280033ef66d06bee;e6d18fff-c2c4-4816-968d-3aed825c18d9) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10453,7 +10421,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8ef9-62496f1b29fdb15063f71d81;a3fded07-ade1-4df0-b645-6d30faee56a5) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e7a3-5b2e63e27aa206d53a733b92;f9dd0a10-b927-478c-bc56-3d19b5d668d7) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10637,7 +10605,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666f87f4-670c0b9e1b477f7e48d59719;bf23f1a5-8d98-4f87-aad3-d52d20cd9514) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d7fa-0482150f770ebb4b6e74d178;223b00c6-129b-4af7-b9fd-7e73bb3db200) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -10726,7 +10694,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8ea0-74aaf60460955f0a0e4f4dc4;efa1bbd1-cfcd-4185-8c19-ec5a5473fdd5) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e752-6edc61576f4f2a535a6c2ba5;059d2da6-1a4a-43c5-b98f-50ad96f3c15d) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10786,7 +10754,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666f87c1-3ca0f09f25c74bc44a626f35;12968414-5658-4fbc-994a-5a5ea6ac5838) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d7d2-6bd7057b110d2a6b311f683f;5cc5dc41-f0c5-4787-b66a-67430a31236c) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -11172,7 +11140,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8e1c-7796863c4dc8de8653602dae;3286fe3b-e471-4116-81a5-13e3119999f4) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e6d9-6c55cb31181ff756705783ee;7173ed94-3724-4ac5-813b-e266ca3923cd) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11288,7 +11256,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8d26-55f3c5ce4045604b00adfaaa;3e8856d2-c83c-4c94-adf2-afd4b856d128) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e610-0bc5abf654a63df43dd35187;84e75c09-b8b7-48ec-8ec1-e4313c2c859b) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11432,7 +11400,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8f7e-45fbfa1f480756c47c9bd90b;83e89b14-93fe-417d-98f7-ba439d3ae119) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e823-2d99ea970aef1cff7f710b90;aa7b0fb2-a68e-4d60-ae6d-fd10739b34b6) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11503,7 +11471,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8cf6-01ac0675779a612b55292acb;65f801bd-2a8b-4cae-99ce-9e9ff7d38d93) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e5e8-1b52fd9d58f8cd06510c08eb;685ff270-0bd6-4030-9473-106ce84c6c3b) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11574,7 +11542,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8dc2-38cb268637a639c63b5c8219;a7ca5a8e-e88a-4096-b061-f8fc3506cb73) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e688-2bfba3af72305086635af116;b42568a9-645d-4ab3-a0c6-df5293384278) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11693,7 +11661,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666f9137-2ad84f3c491b48ee65a6e2ba;e299bd75-1c33-4008-935b-4b7d00adef26) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e9cf-01f023a0275ef861793195a6;0bc50725-bd21-4443-a57c-e11b4a8cc901) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -11778,7 +11746,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8df1-12a365be3eddeb5778087f40;87ef90cf-2a47-411a-b050-91637beec3ad) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e6b0-03cedae11d37b5673dbab861;2edaf4a2-712f-4169-946e-28ff44a621e1) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11840,7 +11808,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666f878c-726c42a822d273275c07dbc8;eb4d1892-e1d0-464a-8d6b-eea2119b8ac9) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d7aa-065b6ae039df0126778fcfcf;067a0fa5-28cb-44e7-8a3e-c30c8ef29229) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -11967,7 +11935,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8d92-126a681c5b6f2ddf7b945c1e;374d4893-e35f-474f-a785-f71600034bab) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e661-48e54e1a55db658e15a18c1d;c038020e-17e9-4f7d-9fb1-f35960ee1f90) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12040,7 +12008,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f881e-582d0c355399ba2f0e5b1e82;5885cbaf-c717-493b-b862-208ab8c38761) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670d823-2ca90f5f4bd0d03d4515874a;07c49e87-ab5e-4d89-9be7-c641fba88c00) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12113,7 +12081,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8f27-7e27688871d8d1bc3c6849d1;37f8f993-841d-400d-8683-a3122cc0bf91) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e7cb-5347d5b011b0448f5bc59179;21e18ee0-4699-42eb-948b-977319fab2ac) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12241,7 +12209,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8b75-54cf34ae41b91d933fb2ba12;4aea5b90-db5d-4968-878c-6151ad9621ca) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e478-5008a0d47709cc06244e826e;da46e1ec-b7f4-4915-a457-4f9fc6a81604) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12435,7 +12403,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8b21-021e164047b9ab1316040df3;a8059565-80c1-44bb-8206-b5afb5be0975) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e41b-6f9412b7463a988d4fb75140;38a752fe-8d90-467f-bc23-969471e3c647) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12505,7 +12473,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f9000-49d560196b5ff41743407a0c;dd2592e3-c408-45b1-bba8-51efd9281380) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e8bf-596d81a648af4e5350f541ef;8217c7b0-ec92-4de5-a807-7e1097c6d6ca) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12575,7 +12543,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8fa6-568990623eda2d5d6ea51f50;bf683cff-378a-48ab-aa1f-cc8140b25cba) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e854-46e6275a1e459ef240a260a8;8247c021-deb4-427a-ad27-eb1100014ceb) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12675,7 +12643,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8e83-6b3914960148a3a42f6608b7;2b04d7d4-8f80-465d-952c-be7a4723f19d) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e737-78b0544b678676b768acd183;f899e8e1-9922-4b53-8d5e-aa727ad43e0b) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12774,7 +12742,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8f06-45f284ee766fb7d91a90387f;0cedeae9-2b62-405e-9f54-18df6033f133) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e7b0-03453e5d712085645d23dec2;e766bff3-194e-4870-bcf8-c0a05b9af713) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -13042,7 +13010,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666f8801-315588f00941d3537505b2fe;b971beb8-a3b9-45af-906a-715d25b03da6) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d808-531128e44dbb7754619cc965;0376f505-b035-49c6-a018-eb146468608b) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -13247,7 +13215,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8ead-0fcd88e96728c3524d398c1b;ea5eb3ed-9024-4875-80eb-0db0f56cddbd) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e75f-63cdeb526ca9cdca3c444a94;f30fc67e-bdaf-4148-9a73-69f4a3cd471d) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -13307,7 +13275,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666f87d0-797bf14914e8fc5b24495627;4326a9fb-ae9d-4e3c-848b-7a8f0a8a72c7) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d7df-0552cfd064e4b79025cbccf7;3bcb6f9b-97ea-4f1a-962b-cdfbedff49f4) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -13832,7 +13800,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666f8542-2dd07a2c748f816c5c7d1c83;01062409-1e80-4c92-9b6a-6e882c2855cf) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d646-7621afd86a2b95f83ab325a5;401d3be0-a311-4b8f-8e54-c06cc59c267c) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -13917,7 +13885,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8e2b-744ed84374b546cd5b716b9c;cad00e94-91c4-4177-a27e-2f1f0e5572bd) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e6e7-4122741b3d6c6f363be61fd2;efe7af33-ef23-45f2-ab45-e72d18806167) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -14061,7 +14029,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8d34-6fd7ee215af80485354b7a23;ef93c58f-f3a4-4b97-b7d5-5508401128f9) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e61d-41adde0d35cdc10261d99ce7;9656a3de-abcf-441e-af33-d4b09ac8fef1) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -14233,7 +14201,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8f8b-36d996b359521003163f78b4;a529288a-b26a-4f28-a31d-e958245d87ca) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e835-57e10b621862fe912d281c13;69a0facf-a74f-4434-a12a-fc9e2e6b3cfc) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -14332,7 +14300,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8d08-162ebee01c9f613c0b96176a;1473e68b-6d4e-4074-b5c9-125af9e95c4f) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e5f5-083264596d0e3ec40622a860;3874cb2b-805c-4c7e-b72e-0567e1804fe0) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -14431,7 +14399,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8dd0-420dec08733a167145343bd1;955a8add-1e0a-4a18-b82e-41fc774e0349) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e696-2f5e48224849abd1519e1557;a0adae52-d2ab-4e2b-a4d6-1824ee7bfa70) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -14522,7 +14490,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666f914b-6fb59c6e7ec3225b5f925965;ddbdd5bd-6937-453d-9b7a-15b17865301a) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e9dc-1a8a06ac1d199a5c037d126a;0b07d4cb-4c47-4783-8401-154595518474) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -14607,7 +14575,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8dff-7f72bbba641bca6465991882;0a3640c6-4c00-4572-b140-72616cbc5e64) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e6be-2d88d6e27045eb320831da3c;29df8040-dd8f-4996-a6f6-83c8293a7735) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -14697,7 +14665,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-666f879a-71fb221162bcdd864af8fa4e;46bcf8fb-e173-484d-98b1-3e7a55b5a7b5) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d7b7-4b55968b0954139166a7ccd3;f15e8e98-b4f7-4fc0-a8fe-645496e4e12c) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -14880,7 +14848,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8da4-6dfb465b392adeb815ecebec;7f950d35-c235-4a5d-912d-949dd4bae873) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e66e-5cfd3fe16d0c476a7eadd055;45969fdb-02cd-4098-b806-cad2e5dd0ef6) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -15009,7 +14977,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f882c-4b9405de5547998111fc8cd2;84e25cb4-9998-4499-983e-3cf133be301d) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670d830-622cd6eb66bca9f53ea1d681;3409feb9-9829-4acc-b737-826f86b44261) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -15138,7 +15106,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8f34-76c037b77ac5a4b93a4b963d;8b0c8e83-b0a2-456a-9ead-4490a90f22d8) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e7d9-5a2f378f3dbd23eb74e5010a;6968c0be-c00f-4f10-9a7a-91234b5da131) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -15238,7 +15206,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8b89-4d50abd72044ecef2e4442ba;68edec1d-76ec-411c-b1c9-94ca78cdc8b6) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e487-2913b802411fa49575848a62;893f7ce9-0012-40ed-87ea-5a921d7ec76c) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -15547,39 +15515,7 @@ ValueError: OPTForCausalLM does not support an attention implementation through ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,microsoft/phi-1_5,microsoft/phi-1_5,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1456.152576,2078.80192,0.0,1432.354816,1207.615488,s,10,0.6942416763305664,0.06942416763305664,0.0002826483816436421,0.06932372665405273,0.0699582633972168,0.06997479438781738,0.06998801918029786,"[0.06995458984375, 0.06928585815429687, 0.0693096923828125, 0.06925567626953125, 0.06919833374023437, 0.06938658905029296, 0.06936601257324218, 0.06915583801269531, 0.06933776092529297, 0.06999132537841797]",tokens/s,3687.476691879045,kWh,8.180421956197512e-07,4.4824246452882447e-07,3.550786365530986e-06,4.817071025679561e-06,tokens/kWh,53144327.46274178,MB,1456.152576,2078.80192,0.0,1432.354816,1306.320384,s,10,36.40508715820313,3.640508715820313,0.0005274391602283375,3.6403291015625,3.641438623046875,3.6414876220703127,3.641526821289063,"[3.641427734375, 3.640418701171875, 3.64153662109375, 3.64002880859375, 3.640720703125, 3.64027978515625, 3.64037841796875, 3.64021435546875, 3.6400849609375, 3.6399970703125]",tokens/s,17.30527377292771,kWh,4.298534585027402e-05,2.3558231379072447e-05,0.0001881406725430684,0.0002546842497724149,tokens/kWh,247365.1199722661,,s,629,36.971170803070116,0.05877769603031807,0.008264322462313465,0.05777305603027344,0.05786112518310547,0.0579207160949707,0.1272486703491211,"[0.057801727294921876, 0.05779455947875976, 0.05784476852416992, 0.0577371826171875, 0.05775462341308594, 0.057778175354003904, 0.05780582427978516, 0.057826305389404295, 0.05805055999755859, 0.05801267242431641, 0.057919486999511716, 0.05777407836914063, 0.057791519165039065, 0.05775971221923828, 0.057834495544433595, 0.05772083282470703, 0.05795942306518555, 0.05778124618530273, 0.05780889511108398, 0.057761791229248044, 0.057766910552978515, 0.05773926544189453, 0.05785599899291992, 0.05767679977416992, 0.05774950408935547, 0.0577894401550293, 0.05772288131713867, 0.05773004913330078, 0.057734142303466796, 0.05782220840454102, 0.05789593505859375, 0.057791488647460934, 0.05770342254638672, 0.057747455596923826, 0.05778739166259766, 0.057734142303466796, 0.05775769424438477, 0.0577710075378418, 0.05780377578735352, 0.057750526428222655, 0.05769420623779297, 0.057829376220703124, 0.0577597770690918, 0.05776892852783203, 0.057734142303466796, 0.05783654403686524, 0.057812992095947265, 0.05779251098632812, 0.05776588821411133, 0.05779046249389649, 0.05778841781616211, 0.05780275344848633, 0.057747455596923826, 0.05777203369140625, 0.05782015991210938, 0.0576993293762207, 0.0577710075378418, 0.057751583099365233, 0.05784675216674805, 0.05773721694946289, 0.057763839721679686, 0.057766910552978515, 0.12727091217041014, 0.05771980667114258, 0.057812992095947265, 0.05775155258178711, 0.05777920150756836, 0.05777407836914063, 0.057728000640869144, 0.05774131011962891, 0.057729022979736325, 0.057850879669189455, 0.057708545684814455, 0.05775564956665039, 0.05777612686157227, 0.05793075180053711, 0.05778432083129883, 0.05775667190551758, 0.057995262145996096, 0.05790822219848633, 0.05772390365600586, 0.05773823928833008, 0.05773926544189453, 0.057903102874755856, 0.057690113067626954, 0.057766910552978515, 0.05768499374389648, 0.05779251098632812, 0.05772185516357422, 0.05770137786865234, 0.05770444869995117, 0.05782732772827148, 0.05771059036254883, 0.05773516845703125, 0.05772697448730469, 0.05782732772827148, 0.05774233627319336, 0.057850879669189455, 0.05775769424438477, 0.057875457763671874, 0.057734142303466796, 0.05773516845703125, 0.057734142303466796, 0.05784985733032227, 0.0577894401550293, 0.05771673583984375, 0.05779558563232422, 0.05785190582275391, 0.05774335861206055, 0.05792768096923828, 0.05777203369140625, 0.05795430374145508, 0.05774335861206055, 0.05771980667114258, 0.05776998519897461, 0.05787136077880859, 0.05775872039794922, 0.05774643325805664, 0.057690113067626954, 0.05787955093383789, 0.05774131011962891, 0.05780582427978516, 0.05769830322265625, 0.057981983184814456, 0.05773513412475586, 0.1272627182006836, 0.05775769424438477, 0.057780223846435545, 0.05795328140258789, 0.05781094360351562, 0.05779455947875976, 0.05775360107421875, 0.05782735824584961, 0.0578109130859375, 0.05780582427978516, 0.05772390365600586, 0.05774028778076172, 0.05779046249389649, 0.05792870330810547, 0.05771878433227539, 0.057760768890380856, 0.05780582427978516, 0.057823230743408206, 0.057867263793945314, 0.057768959045410156, 0.05777305603027344, 0.05779455947875976, 0.05780684661865235, 0.05790617752075195, 0.05782732772827148, 0.05771980667114258, 0.05781913757324219, 0.05770751953125, 0.05773926544189453, 0.05770137786865234, 0.057842689514160155, 0.05772390365600586, 0.05824409484863281, 0.05782220840454102, 0.05778432083129883, 0.05778227233886719, 0.05771059036254883, 0.05783347320556641, 0.05777510452270508, 0.057985023498535154, 0.05771980667114258, 0.057728000640869144, 0.05777920150756836, 0.05775667190551758, 0.057799678802490234, 0.0577259521484375, 0.057829376220703124, 0.057818111419677735, 0.05778841781616211, 0.05778841781616211, 0.057834495544433595, 0.057777153015136716, 0.05777612686157227, 0.057734142303466796, 0.057750526428222655, 0.05788159942626953, 0.05778435134887695, 0.05790000152587891, 0.057783294677734375, 0.05785190582275391, 0.05774028778076172, 0.057777153015136716, 0.05775667190551758, 0.12727603149414063, 0.05770240020751953, 0.057766910552978515, 0.05766963195800781, 0.05784985733032227, 0.057809921264648435, 0.05774643325805664, 0.05775667190551758, 0.05775769424438477, 0.057852928161621096, 0.0577259521484375, 0.057761791229248044, 0.057801727294921876, 0.05786316680908203, 0.057706497192382813, 0.05775667190551758, 0.05778636932373047, 0.05784985733032227, 0.05776998519897461, 0.057717758178710936, 0.057821182250976565, 0.05784473419189453, 0.057673728942871094, 0.05775360107421875, 0.05773311996459961, 0.05784985733032227, 0.05768703842163086, 0.057747455596923826, 0.05777510452270508, 0.057796607971191405, 0.05774028778076172, 0.05779251098632812, 0.05777305603027344, 0.05774643325805664, 0.05769420623779297, 0.05768294525146484, 0.057815040588378906, 0.05780684661865235, 0.05777920150756836, 0.05772697448730469, 0.05780684661865235, 0.057783294677734375, 0.057766910552978515, 0.05775155258178711, 0.05780684661865235, 0.057903102874755856, 0.05777407836914063, 0.05784678268432617, 0.057804798126220705, 0.057876510620117186, 0.05773616027832031, 0.05774848175048828, 0.05779558563232422, 0.05783552169799805, 0.05774643325805664, 0.05774233627319336, 0.05773823928833008, 0.05782527923583984, 0.05770342254638672, 0.057816062927246094, 0.057799678802490234, 0.05786624145507813, 0.05769420623779297, 0.12719821166992187, 0.05774233627319336, 0.05777510452270508, 0.057780223846435545, 0.05775360107421875, 0.05779455947875976, 0.05783859252929688, 0.05774848175048828, 0.057815040588378906, 0.05780070495605469, 0.057729022979736325, 0.05770342254638672, 0.057783294677734375, 0.05780684661865235, 0.05768396759033203, 0.057760768890380856, 0.057777153015136716, 0.057899009704589846, 0.05773209762573242, 0.05783039855957031, 0.05773209762573242, 0.05787136077880859, 0.057659393310546876, 0.05773926544189453, 0.05771263885498047, 0.05778432083129883, 0.0577710075378418, 0.05767270278930664, 0.05772288131713867, 0.057717758178710936, 0.057913345336914064, 0.057852928161621096, 0.05776486587524414, 0.05780377578735352, 0.057809921264648435, 0.05775462341308594, 0.057747455596923826, 0.05783039855957031, 0.05779865646362305, 0.05774233627319336, 0.057744384765625, 0.057747455596923826, 0.05794713592529297, 0.057711616516113284, 0.057834495544433595, 0.05775360107421875, 0.057864192962646485, 0.058229759216308595, 0.057804798126220705, 0.057852928161621096, 0.057839614868164066, 0.05770240020751953, 0.05775564956665039, 0.057818111419677735, 0.05782732772827148, 0.057807872772216794, 0.05770137786865234, 0.057799678802490234, 0.057804798126220705, 0.05776793670654297, 0.05771878433227539, 0.05778124618530273, 0.057785343170166016, 0.12742963409423827, 0.05774028778076172, 0.057807872772216794, 0.05771980667114258, 0.058038272857666016, 0.05782732772827148, 0.05773721694946289, 0.05773311996459961, 0.057747455596923826, 0.057842689514160155, 0.057747455596923826, 0.057763839721679686, 0.05779763031005859, 0.05790719985961914, 0.057760768890380856, 0.05773823928833008, 0.05775667190551758, 0.057852928161621096, 0.05779865646362305, 0.05774643325805664, 0.05780582427978516, 0.05784678268432617, 0.057747455596923826, 0.057744384765625, 0.05780070495605469, 0.05786115264892578, 0.05777916717529297, 0.057695232391357425, 0.057731071472167966, 0.05777407836914063, 0.057729022979736325, 0.05783347320556641, 0.05778739166259766, 0.05775769424438477, 0.0576890869140625, 0.05775360107421875, 0.05779558563232422, 0.057829376220703124, 0.05778432083129883, 0.05771468734741211, 0.057785343170166016, 0.057763839721679686, 0.0577525749206543, 0.05766451263427735, 0.057818111419677735, 0.057796607971191405, 0.05777920150756836, 0.057815040588378906, 0.05781196975708008, 0.05783859252929688, 0.05777612686157227, 0.05771366500854492, 0.05780070495605469, 0.057809921264648435, 0.057785343170166016, 0.05771673583984375, 0.05773004913330078, 0.057799678802490234, 0.05773823928833008, 0.05775155258178711, 0.05774643325805664, 0.057839614868164066, 0.05771366500854492, 0.1273507843017578, 0.05779558563232422, 0.057785343170166016, 0.05778435134887695, 0.05778633499145508, 0.057831424713134766, 0.057826305389404295, 0.057793537139892576, 0.05775360107421875, 0.057821182250976565, 0.05774335861206055, 0.057852928161621096, 0.05775667190551758, 0.05792153549194336, 0.05772390365600586, 0.057760768890380856, 0.0577710075378418, 0.05793894577026367, 0.057728000640869144, 0.05783654403686524, 0.05774335861206055, 0.057903102874755856, 0.057659393310546876, 0.057750526428222655, 0.05767475128173828, 0.057791488647460934, 0.05771980667114258, 0.05774643325805664, 0.05778636932373047, 0.05775667190551758, 0.05774335861206055, 0.05771468734741211, 0.057816062927246094, 0.05776998519897461, 0.057783294677734375, 0.05772803115844727, 0.057756641387939456, 0.05781094360351562, 0.05780582427978516, 0.05770547103881836, 0.05786111831665039, 0.05774643325805664, 0.05785190582275391, 0.05772390365600586, 0.057791488647460934, 0.057760768890380856, 0.05786111831665039, 0.057768959045410156, 0.057766910552978515, 0.057804832458496096, 0.05779555130004883, 0.05770342254638672, 0.057744384765625, 0.05782732772827148, 0.057785343170166016, 0.05776998519897461, 0.05772697448730469, 0.057826305389404295, 0.057777153015136716, 0.05779455947875976, 0.057728000640869144, 0.05781913757324219, 0.05781094360351562, 0.1272125473022461, 0.05771673583984375, 0.057842689514160155, 0.05772390365600586, 0.05778227233886719, 0.05780377578735352, 0.057766910552978515, 0.057766910552978515, 0.057777153015136716, 0.05783039855957031, 0.05773926544189453, 0.05776793670654297, 0.05775667190551758, 0.057818111419677735, 0.05771366500854492, 0.057747455596923826, 0.057807872772216794, 0.057834495544433595, 0.05781708908081055, 0.057760768890380856, 0.05775667190551758, 0.057852928161621096, 0.0576890869140625, 0.05780275344848633, 0.057760768890380856, 0.05777305603027344, 0.057717758178710936, 0.05766656112670898, 0.057834495544433595, 0.05774643325805664, 0.05774950408935547, 0.05773721694946289, 0.05780377578735352, 0.057768959045410156, 0.05773311996459961, 0.057695232391357425, 0.05779251098632812, 0.057777153015136716, 0.05784371185302734, 0.05774335861206055, 0.05781913757324219, 0.05777612686157227, 0.0577628173828125, 0.0576808967590332, 0.05782527923583984, 0.057829376220703124, 0.05774643325805664, 0.057729022979736325, 0.05783859252929688, 0.05792768096923828, 0.057780223846435545, 0.05770751953125, 0.05782220840454102, 0.05778841781616211, 0.05773721694946289, 0.0576993293762207, 0.0577525749206543, 0.05795635223388672, 0.05787443161010742, 0.05776998519897461, 0.05777407836914063, 0.05783654403686524, 0.05774028778076172, 0.12734464263916015, 0.05772185516357422, 0.05777612686157227, 0.05779046249389649, 0.05782527923583984, 0.057899009704589846, 0.057747455596923826, 0.05771878433227539, 0.057731071472167966, 0.05779046249389649, 0.05772083282470703, 0.058065921783447265, 0.05780684661865235, 0.057862144470214844, 0.05772185516357422, 0.05778227233886719, 0.05775974273681641, 0.05793075180053711, 0.057728000640869144, 0.05775974273681641, 0.05777203369140625, 0.0578600959777832, 0.057657344818115235, 0.05773209762573242, 0.05768601608276367, 0.05778636932373047, 0.057729022979736325, 0.0576808967590332, 0.05771059036254883, 0.05782220840454102, 0.05770547103881836, 0.05816524887084961, 0.05773209762573242, 0.057845760345458984, 0.0577525749206543, 0.057673728942871094, 0.05772288131713867, 0.057821182250976565, 0.057695232391357425, 0.05770652770996094, 0.057754592895507814, 0.0577894401550293, 0.0577525749206543, 0.05770547103881836, 0.057745407104492184, 0.057818111419677735, 0.05775360107421875, 0.05780377578735352, 0.0577628173828125, 0.057885696411132816, 0.057777153015136716, 0.05776793670654297, 0.057785343170166016, 0.05783244705200195, 0.05771468734741211, 0.05775360107421875, 0.05771059036254883, 0.05787648010253906, 0.05769728088378906, 0.0577525749206543, 0.05775155258178711, 0.05785190582275391, 0.057678848266601565, 0.12796927642822264, 0.05773823928833008, 0.05773926544189453, 0.057766910552978515, 0.05777612686157227, 0.057826305389404295, 0.05775564956665039, 0.057777153015136716, 0.05773311996459961, 0.05780889511108398, 0.05771980667114258, 0.05772288131713867, 0.05775667190551758, 0.05782732772827148, 0.05779046249389649, 0.05774950408935547, 0.05779763031005859, 0.05782527923583984, 0.057845760345458984, 0.05808127975463867, 0.05776998519897461, 0.0578416633605957, 0.057791488647460934, 0.05775769424438477, 0.05767987060546875, 0.057724929809570315, 0.0578682861328125, 0.05771369552612305, 0.05769315338134766, 0.05768191909790039, 0.057812992095947265, 0.05774028778076172, 0.05774028778076172, 0.05778227233886719, 0.057807872772216794, 0.05773311996459961, 0.05773209762573242, 0.05771263885498047, 0.057867263793945314, 0.057780223846435545, 0.057747455596923826, 0.05771263885498047, 0.05778227233886719, 0.057763839721679686, 0.05777407836914063, 0.05769728088378906, 0.05781196975708008, 0.057899009704589846, 0.05774643325805664, 0.05775564956665039, 0.05781913757324219, 0.05778636932373047, 0.057799678802490234, 0.05776588821411133, 0.057750526428222655, 0.057902080535888675, 0.05771164703369141, 0.05775049591064453, 0.057747455596923826, 0.05784371185302734, 0.0577147216796875, 0.057740257263183596, 0.057766910552978515]",tokens/s,17.0132561760194,, -4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,internlm/internlm2-20b,internlm/internlm2-20b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): - File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch - benchmark_report = Benchmark.launch(benchmark_config) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 47, in launch - report = launcher.launch(worker=cls.run, worker_args=[config]) - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 66, in launch - raise ChildProcessError(response[""traceback""]) -ChildProcessError: Traceback (most recent call last): - File ""/workspace/optimum_benchmark/launchers/process/launcher.py"", line 102, in target - report = worker(*worker_args) - File ""/workspace/optimum_benchmark/benchmark/base.py"", line 60, in run - backend: Backend = backend_factory(backend_config) - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 79, in __init__ - self.load_model_with_no_weights() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 255, in load_model_with_no_weights - self.load_model_from_pretrained() - File ""/workspace/optimum_benchmark/backends/pytorch/backend.py"", line 169, in load_model_from_pretrained - self.pretrained_model = self.automodel_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/models/auto/auto_factory.py"", line 558, in from_pretrained - return model_class.from_pretrained( - File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 3626, in from_pretrained - model = cls(config, *model_args, **model_kwargs) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/f363ea8a116b3ea829c7a068ca24bc9d3e668083/modeling_internlm2.py"", line 974, in __init__ - self.model = InternLM2Model(config) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/f363ea8a116b3ea829c7a068ca24bc9d3e668083/modeling_internlm2.py"", line 796, in __init__ - self.layers = nn.ModuleList([InternLM2DecoderLayer(config) for _ in range(config.num_hidden_layers)]) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/f363ea8a116b3ea829c7a068ca24bc9d3e668083/modeling_internlm2.py"", line 796, in - self.layers = nn.ModuleList([InternLM2DecoderLayer(config) for _ in range(config.num_hidden_layers)]) - File ""/root/.cache/huggingface/modules/transformers_modules/internlm/internlm2-20b/f363ea8a116b3ea829c7a068ca24bc9d3e668083/modeling_internlm2.py"", line 598, in __init__ - self.attention = INTERNLM2_ATTENTION_CLASSES[config.attn_implementation](config=config) -KeyError: 'sdpa' - -",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, +4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,internlm/internlm2-20b,internlm/internlm2-20b,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,,,,,MB,3095.433216,13793.492992,0.0,13147.045888,12781.924864,s,10,8.988890075683594,0.8988890075683594,0.001423213612699,0.8981753845214844,0.9012999267578126,0.9014743347167968,0.9016138610839843,"[0.8981488037109375, 0.8976058959960938, 0.8978265991210937, 0.8977764282226562, 0.8990162353515625, 0.8982019653320312, 0.899675048828125, 0.9012611694335938, 0.9016487426757812, 0.8977291870117188]",tokens/s,284.79600689802794,kWh,1.0606916911072201e-05,5.812208237193772e-06,5.517099321083324e-05,7.159011835909921e-05,tokens/kWh,3575912.5123371445,MB,3096.809472,13793.492992,0.0,13147.045888,12906.706432,s,10,496.00179296875,49.600179296875,0.0025068761942268125,49.599537109375,49.60291796875,49.60425390625,49.60532265625,"[49.60262109375, 49.60017578125, 49.599390625, 49.5969921875, 49.60200390625, 49.59917578125, 49.59676953125, 49.5996796875, 49.60558984375, 49.59939453125]",tokens/s,1.2701566988885713,kWh,0.0005855437311530114,0.00032092896961866243,0.0030658153091317657,0.003972288009903439,tokens/kWh,15859.877189904828,,s,629,503.3055490112303,0.8001678044693649,0.10686695302898579,0.78727783203125,0.7875329833984375,0.7876315185546875,1.6862559765625,"[0.7872880859375, 0.7870259399414062, 0.7873024291992188, 0.787430419921875, 0.7872225341796875, 0.7870167236328125, 0.7872471313476562, 0.787135498046875, 0.7874212036132813, 0.787646484375, 0.7871190795898437, 0.787198974609375, 0.78721533203125, 0.7873003540039063, 0.7873402709960937, 0.7873802490234375, 0.7873699951171875, 0.7872757568359375, 0.7873751220703125, 0.7871262817382813, 0.7870422973632812, 0.78744677734375, 0.7872112426757812, 0.7874457397460938, 0.7872388916015625, 0.7871692504882812, 0.7872634887695312, 0.7872112426757812, 0.7875809326171875, 0.7871897583007812, 0.7871918334960938, 0.7872716674804687, 0.7872532348632812, 0.7870771484375, 0.787178466796875, 0.7871795043945312, 0.7873187866210938, 0.7872747802734374, 0.7873935546875, 0.7871959228515625, 0.787209228515625, 0.7876884765625, 0.7873280029296875, 0.7873853149414063, 0.7873720092773437, 0.787345458984375, 0.7874744262695312, 0.787357666015625, 0.7874508666992187, 0.7872317504882812, 0.78719384765625, 0.7873925170898437, 0.7873443603515625, 0.787577880859375, 0.7875277099609375, 0.7875594482421875, 0.7873863525390625, 0.7875327758789062, 0.7873659057617187, 0.7875850219726562, 0.7875235595703125, 0.787483642578125, 1.68622802734375, 0.7872767944335938, 0.7871539916992187, 0.7875859985351562, 0.7870955810546875, 0.787146728515625, 0.7872962646484375, 0.787230712890625, 0.7869522094726562, 0.7872296752929687, 0.7870853271484375, 0.7870474243164063, 0.7871734008789063, 0.7869942016601562, 0.787184814453125, 0.7870564575195312, 0.7873433837890625, 0.7872163696289063, 0.7871109008789062, 0.7871221923828124, 0.7873392944335937, 0.7873372192382813, 0.7872645263671875, 0.7872798461914062, 0.7871262817382813, 0.7875082397460937, 0.7871580200195313, 0.7872501831054688, 0.7874375610351563, 0.787314697265625, 0.7873116455078125, 0.7874232177734375, 0.78718359375, 0.7870259399414062, 0.7873351440429688, 0.7873904418945312, 0.78729931640625, 0.7872317504882812, 0.7870996704101563, 0.7874580688476562, 0.7870596923828125, 0.7873372192382813, 0.78745703125, 0.7872686157226563, 0.7873843383789062, 0.787409912109375, 0.7872471313476562, 0.7873372192382813, 0.787252197265625, 0.787504150390625, 0.7873597412109375, 0.787330078125, 0.7872645263671875, 0.7876270141601562, 0.78727783203125, 0.7873699951171875, 0.78739453125, 0.7873607788085938, 0.7875860595703125, 0.7873536376953125, 0.78723681640625, 0.7874774780273438, 0.7876034545898437, 1.686560791015625, 0.7868671875, 0.7871846313476563, 0.7875133666992188, 0.78708837890625, 0.78710986328125, 0.7871549682617187, 0.7870218505859375, 0.7871795043945312, 0.7870945434570312, 0.787240966796875, 0.7870156860351563, 0.7869183959960937, 0.7871447143554687, 0.78708837890625, 0.7876351928710937, 0.78715185546875, 0.7872379150390625, 0.7870535888671875, 0.7870084838867187, 0.7871365356445312, 0.78715185546875, 0.7871734619140625, 0.7874733276367187, 0.7874068603515625, 0.7871559448242188, 0.787051513671875, 0.7870802001953126, 0.7871375122070312, 0.7869869995117188, 0.7874959106445313, 0.7874119873046875, 0.7872788696289063, 0.7873699951171875, 0.7870177001953125, 0.7873443603515625, 0.7871549682617187, 0.7873157348632812, 0.7871876831054687, 0.78721435546875, 0.7873106079101563, 0.7871416015625, 0.787336181640625, 0.7874345092773437, 0.7872593994140625, 0.7873280029296875, 0.7874805908203125, 0.787968017578125, 0.7871365356445312, 0.7875297241210938, 0.7873392944335937, 0.7876546630859375, 0.7874345092773437, 0.7874006958007812, 0.787794921875, 0.7873331298828125, 0.7873351440429688, 0.7873392944335937, 0.7871876831054687, 0.7873443603515625, 0.78742529296875, 0.7876710205078125, 0.7873526000976563, 1.6864215087890626, 0.787156982421875, 0.7871692504882812, 0.7870320434570313, 0.7871047973632812, 0.78719384765625, 0.787282958984375, 0.7871682739257813, 0.7870494995117188, 0.7871918334960938, 0.7872501831054688, 0.7871016845703125, 0.7874375610351563, 0.7870842895507812, 0.7874447631835938, 0.7872757568359375, 0.7871201171875, 0.787314697265625, 0.78710888671875, 0.787078125, 0.7875952758789062, 0.7874120483398438, 0.7871282348632812, 0.7870576782226563, 0.7873341674804688, 0.7872553100585937, 0.7871815795898438, 0.7870156860351563, 0.7871897583007812, 0.7872532348632812, 0.7870975952148438, 0.7872348022460938, 0.7872747802734374, 0.7872174072265625, 0.7872041015625, 0.7873034057617188, 0.7870156860351563, 0.787197021484375, 0.7872992553710938, 0.7872368774414062, 0.787378173828125, 0.7874457397460938, 0.787209228515625, 0.7874713745117188, 0.7871426391601563, 0.7871395874023438, 0.7870279541015625, 0.7873341674804688, 0.7872757568359375, 0.7874959106445313, 0.7871610717773437, 0.7873177490234375, 0.7875563354492188, 0.7873013916015625, 0.787205078125, 0.7873659057617187, 0.7874191284179688, 0.7873065185546875, 0.7872808837890625, 0.7878389892578125, 0.7872348022460938, 0.78746826171875, 0.787251220703125, 1.686266845703125, 0.787282958984375, 0.7871907958984375, 0.7871898193359375, 0.7868651123046875, 0.787078125, 0.7870648193359375, 0.7873024291992188, 0.7872276611328125, 0.78713037109375, 0.7872348022460938, 0.7871282958984375, 0.7874324340820312, 0.7871651611328125, 0.7873106079101563, 0.787651611328125, 0.7873771362304688, 0.7877498779296875, 0.7874212036132813, 0.7871027221679687, 0.7872348022460938, 0.7873812255859375, 0.7873751220703125, 0.7872491455078126, 0.787462158203125, 0.7871181030273438, 0.7874293823242188, 0.7872245483398438, 0.7871190795898437, 0.7873648681640625, 0.7871498413085938, 0.7873106079101563, 0.78740478515625, 0.787589111328125, 0.787209228515625, 0.7874078979492187, 0.7869603881835937, 0.787103759765625, 0.7870873413085937, 0.7873751220703125, 0.7874447631835938, 0.787420166015625, 0.787315673828125, 0.7874508666992187, 0.787472412109375, 0.7872890625, 0.7874396362304688, 0.787357666015625, 0.787262451171875, 0.7875082397460937, 0.7873976440429687, 0.7874017333984376, 0.7876085815429688, 0.787541015625, 0.7874969482421875, 0.787430419921875, 0.7876658935546875, 0.7873812255859375, 0.7873597412109375, 0.7874191284179688, 0.7873771362304688, 0.7876802368164062, 0.7875235595703125, 1.6858050537109375, 0.7872235717773437, 0.7872061157226562, 0.7873433837890625, 0.7870330810546875, 0.7872696533203125, 0.787198974609375, 0.787314697265625, 0.7871016845703125, 0.7874959106445313, 0.7869450073242188, 0.787140625, 0.7872020263671875, 0.7870494995117188, 0.7871918334960938, 0.787009521484375, 0.7873853149414063, 0.7873126220703125, 0.7870853271484375, 0.7873423461914062, 0.7871283569335937, 0.7871856079101562, 0.787072021484375, 0.7873239135742187, 0.7870382080078125, 0.7873341674804688, 0.7877058715820312, 0.7871836547851563, 0.7872327270507813, 0.787140625, 0.7872317504882812, 0.78727783203125, 0.7871293334960937, 0.7871600341796875, 0.7873535766601563, 0.7871764526367188, 0.7871395874023438, 0.7871068115234375, 0.7872532348632812, 0.787435546875, 0.7875829467773438, 0.7872645263671875, 0.7874283447265625, 0.7873167114257813, 0.7873587036132812, 0.7876290283203125, 0.7873638305664062, 0.787304443359375, 0.7874037475585938, 0.787282958984375, 0.7874293823242188, 0.7872808837890625, 0.7875972900390625, 0.7874498291015625, 0.7873607788085938, 0.7873402709960937, 0.7875440673828125, 0.7874150390625, 0.7875194702148437, 0.7875552978515625, 0.7872030639648437, 0.7873720092773437, 0.7872184448242188, 1.6870113525390624, 0.7871713256835937, 0.7869183959960937, 0.7871068115234375, 0.7869962158203125, 0.7871795043945312, 0.7872501831054688, 0.7877222290039062, 0.7870320434570313, 0.78708837890625, 0.7871897583007812, 0.7871375122070312, 0.7872225341796875, 0.7870443725585937, 0.7869265747070312, 0.7872819213867187, 0.7871897583007812, 0.7873556518554687, 0.78718359375, 0.7872952270507813, 0.787251220703125, 0.7871815795898438, 0.7872481079101562, 0.7872788696289063, 0.7871242065429688, 0.7873423461914062, 0.7870812377929688, 0.7873597412109375, 0.7870422973632812, 0.7870965576171876, 0.7872737426757812, 0.7871129760742187, 0.7872225341796875, 0.78717236328125, 0.7872665405273438, 0.787083251953125, 0.7871498413085938, 0.7871826171875, 0.7873556518554687, 0.7874365234375, 0.7871846313476563, 0.7872051391601562, 0.787178466796875, 0.7873925170898437, 0.7873955688476563, 0.78745703125, 0.787177490234375, 0.7872665405273438, 0.7871211547851562, 0.78724609375, 0.7872604370117188, 0.78769970703125, 0.7871692504882812, 0.787398681640625, 0.7871006469726562, 0.7874293823242188, 0.787251220703125, 0.7878092651367188, 0.78738330078125, 0.78727783203125, 0.7872849731445313, 0.787525634765625, 0.7875266723632812, 1.688580078125, 0.7875010375976562, 0.78710888671875, 0.787219482421875, 0.78703515625, 0.7871795043945312, 0.7870013427734375, 0.787435546875, 0.7873505249023437, 0.7871016845703125, 0.78717236328125, 0.7871641845703125, 0.7875369262695312, 0.7873310546875, 0.78725732421875, 0.7872532348632812, 0.7872379150390625, 0.7872910766601563, 0.7870904541015625, 0.7873751220703125, 0.7871580810546875, 0.7873095092773438, 0.787483642578125, 0.7870177001953125, 0.7870822143554688, 0.7873853149414063, 0.787145751953125, 0.7872327880859376, 0.78710888671875, 0.7874580688476562, 0.7870371704101562, 0.7873546142578125, 0.7873228759765625, 0.787219482421875, 0.7871498413085938, 0.7874293823242188, 0.7872030639648437, 0.7873638305664062, 0.7872952270507813, 0.7876177978515625, 0.787346435546875, 0.78724609375, 0.787198974609375, 0.7873720092773437, 0.7875277099609375, 0.7873535766601563, 0.7872420043945313, 0.7872880859375, 0.7872399291992187, 0.7871764526367188, 0.787567626953125, 0.787324951171875, 0.7877447509765625, 0.78729931640625, 0.7872399291992187, 0.7876454467773437, 0.787399658203125, 0.7873812255859375, 0.787399658203125, 0.7873167114257813, 0.7872716674804687, 0.7875164184570312, 0.7873894653320312, 1.6889835205078125, 0.78731982421875, 0.7872686157226563, 0.78731982421875, 0.7867996215820312, 0.7874959106445313, 0.7870904541015625, 0.78729931640625, 0.7871846313476563, 0.78740478515625, 0.7873280029296875, 0.7871190795898437, 0.787230712890625, 0.7871426391601563, 0.7872501831054688, 0.7871344604492188, 0.787125244140625, 0.7872696533203125, 0.7874662475585937, 0.78742626953125, 0.787325927734375, 0.7875921630859375, 0.7871027221679687, 0.787114990234375, 0.7877273559570312, 0.7872798461914062, 0.7875399780273438, 0.7872379150390625, 0.7872163696289063, 0.7874652099609375, 0.7871846313476563, 0.787125244140625, 0.7873515625, 0.7874293823242188, 0.7873546142578125, 0.7871641845703125, 0.7871826171875, 0.787314697265625, 0.7872696533203125, 0.7872747802734374, 0.78727783203125, 0.7873239135742187, 0.78721533203125, 0.787272705078125, 0.7872870483398438, 0.7874498291015625, 0.7874037475585938, 0.7873505249023437, 0.7871979370117187, 0.7875717163085938, 0.7872645263671875, 0.7875338134765625, 0.7875123291015625, 0.7872870483398438, 0.787283935546875, 0.7875277099609375, 0.7874907836914062, 0.7879116821289063, 0.7873699951171875, 0.7874068603515625, 0.7874037475585938, 0.7875430297851562, 0.7881441040039062, 1.688532958984375, 0.7873351440429688, 0.78715185546875, 0.7871795043945312, 0.7868784790039063, 0.7871682739257813, 0.7871661987304688, 0.787072998046875, 0.7870259399414062, 0.7871641845703125, 0.7871897583007812, 0.7870986328125, 0.78719384765625, 0.7872757568359375, 0.7872184448242188, 0.7872747802734374, 0.787252197265625, 0.7876331787109375, 0.7872133178710937, 0.7871109008789062, 0.7872471313476562, 0.7872225341796875, 0.7870023803710937, 0.7875215454101563, 0.7872635498046875, 0.7872357788085937, 0.7873863525390625, 0.7871764526367188, 0.7871734008789063, 0.7871651611328125, 0.7871242065429688, 0.787431396484375, 0.7871795043945312, 0.7872102661132813, 0.787019775390625, 0.7872379150390625, 0.7876024169921875, 0.7872655639648437, 0.7871641845703125, 0.7872716674804687, 0.7873771362304688, 0.787409912109375, 0.7874385986328125, 0.7872849731445313, 0.7871764526367188, 0.7874283447265625, 0.7873208618164063, 0.7872880859375, 0.78717236328125, 0.7874662475585937, 0.7873935546875, 0.7875758056640625, 0.78751025390625, 0.7873925170898437, 0.7873597412109375, 0.7873290405273438, 0.7875635375976563, 0.7874447631835938, 0.7875901489257813, 0.7875963134765624, 0.7870873413085937, 0.7879024658203125, 0.7874406127929687]",tokens/s,1.2497378605018417,, 4bit-gptq-exllama-v2-sdpa,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,m,m,cuda,0,42,,,True,True,True,True,,float16,True,False,,sdpa,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): File ""/workspace/llm_perf/update_llm_perf_cuda_pytorch.py"", line 148, in benchmark_cuda_pytorch benchmark_report = Benchmark.launch(benchmark_config) @@ -15619,7 +15555,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8b2e-2007dda5055c71fd524a5a08;21a26664-14cb-4e8d-84fd-12a030bfdce6) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e429-10e69ddb2ecf679c34ff7f2a;a9c4f355-4469-49be-85cc-a444fdb0295d) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -15689,7 +15625,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f900d-302bd90e1da5bc733e310fff;fe20ca26-480f-46cc-ab8e-183f5e1ac7dc) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e8ce-6b75b9030643d3775c4fb118;cfb334d2-3a06-4dd6-a5da-201d4ad139a0) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -15759,7 +15695,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-666f8fb7-555de556340ce01a6fe66dee;8cba574c-2d51-4e9a-83a4-2327f428f951) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e86b-1ba175c27622547e7edc0e86;1e0e9503-aaa5-4f1a-970a-dffa40e3a7bf) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`.