diff --git "a/perf-df-gptq-1xA10.csv" "b/perf-df-gptq-1xA10.csv" --- "a/perf-df-gptq-1xA10.csv" +++ "b/perf-df-gptq-1xA10.csv" @@ -67,7 +67,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e73d-6e78a4de38189e871a8fb8d9;33f1f48a-2eb5-47a5-99ae-2034efbbdf2d) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722f61-17d326461f93f7cc54ba233e;735e4b1d-92d2-4cf3-bef4-b1ee95e7a86c) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -138,7 +138,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e7b6-0dba1752652b817730e4221d;49a3c763-779d-4a5c-aa75-005d5680c464) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722fdc-2ffc22b948ae5038381299f9;06791a61-0966-4e5e-83e4-8a695491b409) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -350,7 +350,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d80f-45283a9e7d8507c45acf2b6c;c2883c49-e2aa-4600-bdbb-f887d35ae28f) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6672299c-22039d011ef110ac456b59a4;e7969b9c-8251-4622-9966-50e7f3991612) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -439,7 +439,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e766-56326525713b95e37f5d2d73;d29bda65-aa76-488a-ae14-b94899cfd606) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722f89-148b07ee47b17d6a0ee309ca;3224aa8b-9e26-4f1d-b1cc-5d40c423a706) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -499,7 +499,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d7e6-2b07a77e5bd821225e3d22b0;30fa1050-f677-4cf8-8f1b-3c1ae4a38d8a) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66722973-492ae14f2922a313349e4c29;e0b3ba1c-65b2-4fba-bc9d-b8352cc5c402) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -731,7 +731,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp0dciw21u/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpr0egzqov/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -828,7 +828,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d64c-3e731a8532df908b7cb6ee4a;faaba032-7046-4954-a6eb-565ec430ccc4) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667227ea-2cf5ee4511d4efd7470e596f;dc1b4b98-b26e-4988-b43a-8cc7c8d91dd0) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -913,7 +913,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e6ed-6111c8325a5e291e3592c6fa;fec846fa-71c7-4c60-8c0c-97daef167832) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722f12-0652db513e41b6b35491d2db;f7f01c3e-f291-44fd-83f2-ce6e4118d4f6) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1029,7 +1029,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e624-5e5f0f20462627de0eeb3134;79dd094f-8cd6-4693-a0bc-7ab8a901ddde) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722e49-3078c7032b4555712f4d838e;7fdd744d-2a22-45fb-850c-aa487a1914ae) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1129,7 +1129,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpqzab82ew/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpb4p_abr2/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -1201,7 +1201,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e83f-0f5844cc62326de7034419c5;f2e8d84d-819c-4f29-bb95-bf337fe739bc) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66723055-331d050a5483e2957cfbd3b1;99fb6a02-33cd-4ca7-a824-1314926aabcb) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1258,7 +1258,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp_51ktme_/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpfsqccglh/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,M,M,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -1300,7 +1300,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e5fc-2c5d453e387404071f4268b4;7a68e059-8daf-43f9-8ae9-570e4070d813) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722e20-42031b194ddaa2836ad246b3;8bb2de05-c67f-45d6-9ad6-11085db43779) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1371,7 +1371,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e69c-507e69c1320f3d3b5ee71618;46fb56ae-bad2-47f3-80ff-f00e9e88be40) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722ec1-7a815b7b5913790805d5d987;8952325c-3f9f-4931-b8a7-e9a66b8ce932) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1429,7 +1429,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp_91tyht6/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp2equ932i/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,4706.717696,21498.42944,0.0,20851.982336,20236.60032,s,10,15.640249145507811,1.5640249145507812,0.00168203754809238,1.563786376953125,1.564931494140625,1.5666584350585937,1.5680399877929687,"[1.5630980224609374, 1.5644747314453125, 1.5625794677734375, 1.56453466796875, 1.5645272216796875, 1.5645477294921875, 1.5624769287109375, 1.5628692626953125, 1.5683853759765625, 1.5627557373046874]",tokens/s,163.68025702041214,kWh,1.8462728808323543e-05,1.0117648402811028e-05,9.129585081439906e-05,0.00011987622802553362,tokens/kWh,2135535.9958896274,MB,4706.717696,21498.42944,0.0,20851.982336,20339.706368,s,10,879.7075468749999,87.97075468749999,0.002247932535373388,87.97129296874999,87.973384375,87.973496875,87.973586875,"[87.9672265625, 87.968984375, 87.9679609375, 87.9719375, 87.972140625, 87.973609375, 87.973359375, 87.968734375, 87.9729453125, 87.9706484375]",tokens/s,0.7161470902892214,kWh,0.0010385013203819595,0.0005691906291757731,0.005147801034904198,0.0067554929844619305,tokens/kWh,9325.744271351337,,s,629,892.387736694337,1.4187404398956058,0.1859067280964402,1.39627001953125,1.3965248535156252,1.3967189697265625,2.9604817578125,"[1.39608984375, 1.3963448486328125, 1.3963612060546875, 1.3961072998046875, 1.3964923095703126, 1.3963653564453125, 1.396220947265625, 1.3965025634765624, 1.3960928955078125, 1.3962034912109376, 1.3961072998046875, 1.396042724609375, 1.396294677734375, 1.3963448486328125, 1.3963970947265625, 1.3963560791015626, 1.39631201171875, 1.396305908203125, 1.3963643798828125, 1.3958726806640624, 1.3961134033203124, 1.3959710693359375, 1.3963284912109375, 1.396166748046875, 1.3961839599609376, 1.3961236572265625, 1.3963602294921875, 1.3963294677734375, 1.3962445068359375, 1.396177001953125, 1.3961810302734374, 1.3960733642578125, 1.3959208984375, 1.396359130859375, 1.3958072509765624, 1.39637451171875, 1.396179931640625, 1.396137939453125, 1.39983154296875, 1.396505615234375, 1.3963970947265625, 1.3963028564453126, 1.396326416015625, 1.3964442138671875, 1.3961175537109376, 1.396173828125, 1.396177978515625, 1.3965987548828125, 1.3961728515625, 1.3964912109375, 1.3962935791015625, 1.39636328125, 1.39614208984375, 1.3962025146484376, 1.396126708984375, 1.3964083251953125, 1.3963018798828124, 1.396030517578125, 1.3959925537109374, 1.39601513671875, 1.395998779296875, 1.39626904296875, 2.96171923828125, 1.396209716796875, 1.3962506103515624, 1.3962557373046875, 1.3960858154296876, 1.3961513671875, 1.39597412109375, 1.39620654296875, 1.396126708984375, 1.396136962890625, 1.3962086181640625, 1.396041748046875, 1.39642578125, 1.396341796875, 1.3963878173828126, 1.396432861328125, 1.3963028564453126, 1.395948486328125, 1.3960714111328125, 1.3977159423828125, 1.395883056640625, 1.3964984130859375, 1.3963028564453126, 1.3964410400390626, 1.3959066162109375, 1.3964769287109375, 1.396209716796875, 1.3961768798828125, 1.39627001953125, 1.396104248046875, 1.396179931640625, 1.396157470703125, 1.39639404296875, 1.3961943359375, 1.39627001953125, 1.3962998046875, 1.39653125, 1.3978797607421876, 1.39601513671875, 1.39614111328125, 1.3961656494140624, 1.3962823486328124, 1.396073486328125, 1.3962823486328124, 1.3959659423828126, 1.3963233642578126, 1.3965946044921875, 1.3960704345703125, 1.396041748046875, 1.3960120849609374, 1.396105224609375, 1.3962178955078124, 1.39605810546875, 1.396042724609375, 1.3964912109375, 1.3962711181640626, 1.396250732421875, 1.396602783203125, 1.3962762451171875, 1.396516845703125, 1.396326416015625, 1.398640625, 1.3963304443359374, 2.96060302734375, 1.39647900390625, 1.39653125, 1.3961614990234374, 1.3960181884765626, 1.3962342529296874, 1.39601611328125, 1.3958338623046875, 1.39621484375, 1.3961512451171876, 1.3961553955078125, 1.3965394287109374, 1.39652294921875, 1.3962301025390624, 1.396216796875, 1.39635400390625, 1.3961707763671876, 1.396279296875, 1.39607861328125, 1.3962249755859375, 1.395962890625, 1.3963018798828124, 1.396126708984375, 1.3988648681640625, 1.3960888671875, 1.3964031982421874, 1.3962301025390624, 1.3964400634765626, 1.3959556884765625, 1.3960714111328125, 1.3961871337890626, 1.396095947265625, 1.3961861572265626, 1.3961195068359376, 1.3960130615234374, 1.395988525390625, 1.3964349365234374, 1.396515869140625, 1.3959833984375, 1.396220947265625, 1.3962230224609375, 1.396178955078125, 1.396315185546875, 1.3963509521484374, 1.3963929443359375, 1.3960755615234375, 1.3961195068359376, 1.3990062255859375, 1.396305908203125, 1.3959075927734375, 1.3962926025390625, 1.3961728515625, 1.3962230224609375, 1.39626806640625, 1.3965035400390624, 1.3963131103515625, 1.3960294189453124, 1.3960611572265624, 1.396111328125, 1.396221923828125, 1.39627734375, 1.3961727294921875, 1.3960447998046874, 2.961617919921875, 1.3961595458984375, 1.39620458984375, 1.39593212890625, 1.3961533203125, 1.3960263671875, 1.3961451416015624, 1.3962352294921876, 1.3962076416015625, 1.3960560302734375, 1.3962998046875, 1.3964493408203125, 1.39588720703125, 1.396454345703125, 1.3963284912109375, 1.39639501953125, 1.3963642578125, 1.3962127685546875, 1.396188232421875, 1.396221923828125, 1.396116455078125, 1.3964298095703125, 1.3964974365234375, 1.39631201171875, 1.3961666259765626, 1.396533203125, 1.3962506103515624, 1.396453369140625, 1.3960457763671874, 1.396447265625, 1.396515869140625, 1.39631005859375, 1.396305908203125, 1.3964073486328126, 1.396190185546875, 1.3960980224609374, 1.39657421875, 1.396528076171875, 1.3965926513671876, 1.4005042724609376, 1.396791259765625, 1.3964676513671874, 1.3963724365234376, 1.3963826904296874, 1.3961553955078125, 1.396410400390625, 1.39633251953125, 1.3959403076171875, 1.396326416015625, 1.3962608642578125, 1.3963551025390626, 1.3964451904296875, 1.396463623046875, 1.39631103515625, 1.39641552734375, 1.3962301025390624, 1.396401123046875, 1.3960284423828124, 1.3962137451171874, 1.3964359130859374, 1.3964512939453124, 1.3963233642578126, 1.39616357421875, 2.96169482421875, 1.39649951171875, 1.39685986328125, 1.396106201171875, 1.3962987060546874, 1.396274169921875, 1.3959752197265625, 1.3962496337890624, 1.3964339599609374, 1.39608984375, 1.3961011962890626, 1.3962977294921874, 1.3963929443359375, 1.39652197265625, 1.3963203125, 1.3962506103515624, 1.396452392578125, 1.3964031982421874, 1.3964390869140626, 1.396611083984375, 1.3960335693359376, 1.3959803466796874, 1.395914794921875, 1.3961964111328125, 1.3962803955078125, 1.3961553955078125, 1.3965179443359375, 1.39654248046875, 1.3962127685546875, 1.3964461669921875, 1.396337646484375, 1.3962884521484376, 1.3965230712890624, 1.3961533203125, 1.39621484375, 1.3961461181640624, 1.3964666748046874, 1.396431884765625, 1.3962578125, 1.396168701171875, 1.3964727783203126, 1.4002913818359375, 1.396400146484375, 1.3963284912109375, 1.3961728515625, 1.396716552734375, 1.396369384765625, 1.3964410400390626, 1.3962874755859376, 1.3961717529296875, 1.39602734375, 1.396189208984375, 1.39620458984375, 1.3962147216796874, 1.3963602294921875, 1.396046875, 1.39650048828125, 1.3963643798828125, 1.396154296875, 1.396696044921875, 1.396347900390625, 1.3963560791015626, 1.3960509033203126, 2.96171630859375, 1.3963919677734375, 1.3963355712890626, 1.39618505859375, 1.3962926025390625, 1.396115478515625, 1.3960120849609374, 1.396125732421875, 1.396279296875, 1.3963253173828125, 1.3961246337890625, 1.3961092529296875, 1.3965772705078126, 1.396653076171875, 1.3962025146484376, 1.395925048828125, 1.39633154296875, 1.396427734375, 1.3964052734375, 1.3962967529296875, 1.39624755859375, 1.396263916015625, 1.3963182373046874, 1.396494384765625, 1.3964830322265624, 1.3963775634765625, 1.396209716796875, 1.396411376953125, 1.3961912841796875, 1.39610107421875, 1.39620556640625, 1.39635205078125, 1.3962015380859376, 1.396042724609375, 1.3961400146484375, 1.396209716796875, 1.3963079833984375, 1.3961953125, 1.40067431640625, 1.3964073486328126, 1.396336669921875, 1.3963826904296874, 1.396295654296875, 1.39614208984375, 1.3964093017578125, 1.396348876953125, 1.39626806640625, 1.39626806640625, 1.3963868408203124, 1.3961544189453126, 1.3960079345703125, 1.39607861328125, 1.396130859375, 1.3962720947265626, 1.396485107421875, 1.39627001953125, 1.3966878662109374, 1.396177978515625, 1.3960714111328125, 1.39645751953125, 1.3967821044921875, 1.3963929443359375, 1.396221923828125, 2.9624052734375, 1.39640625, 1.396473876953125, 1.396378662109375, 1.3961103515625, 1.3961666259765626, 1.3961380615234376, 1.3961973876953124, 1.3964195556640624, 1.3963243408203125, 1.39621484375, 1.39620556640625, 1.3961348876953126, 1.3965609130859375, 1.3961318359375, 1.396334716796875, 1.396403076171875, 1.3967677001953125, 1.3963458251953125, 1.39630078125, 1.396005859375, 1.396190185546875, 1.396221923828125, 1.3991322021484376, 1.396274169921875, 1.3965567626953126, 1.3959332275390626, 1.39675439453125, 1.3960386962890625, 1.396147216796875, 1.3964830322265624, 1.3962813720703124, 1.3961563720703125, 1.3962669677734374, 1.3960653076171874, 1.3961922607421875, 1.3961605224609375, 1.3963775634765625, 1.396552734375, 1.39624853515625, 1.396236328125, 1.396547607421875, 1.396465576171875, 1.3962352294921876, 1.3962802734375, 1.3960560302734375, 1.3962137451171874, 1.39620556640625, 1.396074462890625, 1.396125732421875, 1.396263916015625, 1.3962332763671874, 1.396306884765625, 1.396263916015625, 1.39645849609375, 1.396369384765625, 1.3965506591796875, 1.396136962890625, 1.395937255859375, 1.3961185302734376, 1.3962281494140625, 1.3962301025390624, 1.400385498046875, 2.960169921875, 1.3960386962890625, 1.3961124267578124, 1.3962587890625, 1.3961328125, 1.3968507080078125, 1.396115478515625, 1.39623828125, 1.3961759033203125, 1.396401123046875, 1.39621484375, 1.3962935791015625, 1.39642578125, 1.3964779052734375, 1.3963294677734375, 1.3960775146484374, 1.397832763671875, 1.396600830078125, 1.39633154296875, 1.3963253173828125, 1.3960980224609374, 1.3963857421875, 1.3962301025390624, 1.39618408203125, 1.396242431640625, 1.3963233642578126, 1.3963560791015626, 1.3964493408203125, 1.39633154296875, 1.396285400390625, 1.3962772216796875, 1.3962454833984375, 1.3963140869140624, 1.396157470703125, 1.39630078125, 1.3960755615234375, 1.3961595458984375, 1.39652392578125, 1.396315185546875, 1.3964451904296875, 1.3963427734375, 1.3962332763671874, 1.39616357421875, 1.3960242919921875, 1.3964410400390626, 1.396295654296875, 1.396368408203125, 1.396148193359375, 1.3961932373046875, 1.396262939453125, 1.3962987060546874, 1.3963458251953125, 1.396189208984375, 1.3962659912109374, 1.3962874755859376, 1.3962987060546874, 1.3963837890625, 1.3963499755859374, 1.3961451416015624, 1.396453369140625, 1.3965670166015625, 1.3963551025390626, 1.3961656494140624, 2.9656298828125, 1.3966131591796875, 1.396486083984375, 1.396496337890625, 1.3961390380859375, 1.3962127685546875, 1.396168701171875, 1.39631005859375, 1.3962650146484374, 1.39608984375, 1.396369384765625, 1.396193359375, 1.3961533203125, 1.396337646484375, 1.396303955078125, 1.3962352294921876, 1.3963612060546875, 1.3983529052734376, 1.39636328125, 1.396357177734375, 1.396263916015625, 1.3963140869140624, 1.3959822998046876, 1.396474853515625, 1.396347900390625, 1.396595703125, 1.396111328125, 1.3961328125, 1.3962701416015626, 1.3961102294921874, 1.3961707763671876, 1.39612158203125, 1.3963203125, 1.3962547607421876, 1.396042724609375, 1.396252685546875, 1.3960447998046874, 1.396010986328125, 1.3960858154296876, 1.3964871826171874, 1.396463623046875, 1.396033447265625, 1.3962823486328124, 1.3962874755859376, 1.3961461181640624, 1.396464599609375, 1.3964410400390626, 1.396326416015625, 1.396304931640625, 1.396130859375, 1.3960826416015626, 1.396231201171875, 1.39595166015625, 1.3962086181640625, 1.39643505859375, 1.3962188720703126, 1.39656396484375, 1.396241455078125, 1.396109375, 1.396360107421875, 1.3962454833984375, 1.396515869140625, 1.4013460693359374, 2.9601669921875, 1.396242431640625, 1.3961563720703125, 1.3967093505859376, 1.3961451416015624, 1.3960601806640625, 1.3962076416015625, 1.396126708984375, 1.396243408203125, 1.3964246826171876, 1.3964666748046874, 1.39620654296875, 1.39607958984375, 1.39641552734375, 1.396093994140625, 1.3960858154296876, 1.3963212890625, 1.396093994140625, 1.3965445556640625, 1.39627001953125, 1.3963775634765625, 1.396485107421875, 1.3963131103515625, 1.3964759521484376, 1.3963275146484375, 1.3965045166015626, 1.39883935546875, 1.3962230224609375, 1.3962762451171875, 1.396242431640625, 1.3961400146484375, 1.3962178955078124, 1.3964820556640625, 1.396148193359375, 1.3961688232421876, 1.3961829833984376, 1.396322265625, 1.396190185546875, 1.396115478515625, 1.39614208984375, 1.3964031982421874, 1.39636328125, 1.3966151123046875, 1.3964134521484375, 1.3963355712890626, 1.396177978515625, 1.396125732421875, 1.396169677734375, 1.396275146484375, 1.3965240478515626, 1.3966817626953125, 1.3962578125, 1.396337646484375, 1.3967205810546874, 1.3967247314453124, 1.396284423828125, 1.396473876953125, 1.3963509521484374, 1.3962127685546875, 1.3961492919921874, 1.3961031494140625, 1.3963387451171876, 1.39633251953125]",tokens/s,0.7048505645427167,, @@ -1462,7 +1462,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e9e3-12b9040642b4ce1a06e877f7;b70095d1-5ed2-4513-87dc-e4896b71b955) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667231d0-175e2f0f418e6dbd0637e8fc;4b1dfaf6-5e7b-4732-b4dc-63f04caad243) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -1547,7 +1547,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e6c4-76631f0e6ce673311bbbf2fa;0d057160-c793-4984-becc-f079016e6c5c) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722ee9-53fbd6b5077f9ee30bd3b21a;c66409c2-9004-47e3-8734-919ffcb8b6b2) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1609,7 +1609,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d7be-7f50158c75e32bc36418b187;eb0c410e-0157-4fda-935a-bed30d952bb1) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6672294b-6579e11c403be6ef38e48a60;af4af45b-358d-4221-aa0c-f8ae312508b2) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -1736,7 +1736,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e674-6605acc65b6bdd7e098f8bad;90df0b50-2962-4398-af50-399139f8c444) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722e9a-2d21eaec0c92b09427d88430;2ab68778-2fb2-4041-b15e-749f469261a0) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1809,7 +1809,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670d837-571681051a23960e1e9aefa3;b7def1cb-7d93-455a-a5d3-6807c4f3fd7a) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667229c5-2c0bb0e26fbda039612b2fe8;5ae46a3a-7aec-49fa-a821-4fd3cfebb6f5) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1868,7 +1868,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp53bqdw_e/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmppaf9rv3a/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,B,B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -1910,7 +1910,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e7df-20a3ef7e5b1e2ba96016b916;f1079b5b-4f5e-42cf-be3b-f376e9c2739f) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66723006-24528c0113b23fc90d2283fc;afb826f9-4255-4172-b91b-b4b1f8d3e0fb) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -1967,7 +1967,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp1ascikep/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpibk0v3_c/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,TencentARC/Mistral_Pro_8B_v0.1,TencentARC/Mistral_Pro_8B_v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,5946.503168,7053.246464,0.0,6406.79936,6098.60864,s,10,4.370441192626953,0.4370441192626952,0.0009281353036040856,0.43696878051757815,0.4381058013916016,0.4383167251586914,0.4384854641723633,"[0.4385276489257813, 0.4359717712402344, 0.43771954345703123, 0.43744598388671874, 0.4380589294433594, 0.4362575378417969, 0.43786093139648435, 0.4364915771484375, 0.436230712890625, 0.43587655639648437]",tokens/s,585.7532196792366,kWh,5.148596204997261e-06,2.8211948946868096e-06,2.4996191494520076e-05,3.296598259420415e-05,tokens/kWh,7765580.754902423,MB,5946.503168,7055.343616,0.0,6408.896512,6157.952,s,10,244.94424023437497,24.494424023437496,0.005272769015065994,24.493744140625,24.502037109375,24.502107421875,24.502163671875,"[24.487064453125, 24.502177734375, 24.49926953125, 24.4968515625, 24.502021484375, 24.493955078125, 24.493533203125, 24.4922109375, 24.487486328125, 24.489669921875]",tokens/s,2.5720139383444343,kWh,0.00028916327015789236,0.00015848618370445323,0.00135038218658508,0.0017980316404474258,tokens/kWh,35038.315557296286,,s,629,248.48397949218753,0.3950460723246224,0.05195367916433446,0.388748291015625,0.38915277099609374,0.38942371215820315,0.8254186791992187,"[0.38863052368164064, 0.3885834350585938, 0.38854656982421876, 0.38854452514648435, 0.3885762634277344, 0.3887677307128906, 0.3885189208984375, 0.3887728576660156, 0.3885434875488281, 0.38884658813476564, 0.38842572021484373, 0.3884328918457031, 0.3885711364746094, 0.3883417663574219, 0.3885240173339844, 0.38941387939453126, 0.3884031982421875, 0.3887196044921875, 0.38854452514648435, 0.3887032470703125, 0.3886264343261719, 0.38861721801757815, 0.3888977966308594, 0.3890206604003906, 0.388917236328125, 0.3884503173828125, 0.388843505859375, 0.3885404052734375, 0.38881793212890625, 0.38851788330078124, 0.3882977294921875, 0.3884666748046875, 0.388701171875, 0.38850970458984374, 0.3885455322265625, 0.3883397216796875, 0.3884103698730469, 0.388495361328125, 0.3894988708496094, 0.3882915954589844, 0.3888158569335938, 0.3886602783203125, 0.3886519775390625, 0.3887790222167969, 0.3884134521484375, 0.38843392944335936, 0.3885281372070313, 0.3888619384765625, 0.38980096435546874, 0.38914764404296875, 0.3887196044921875, 0.38860186767578125, 0.38848306274414063, 0.38838375854492185, 0.3885035400390625, 0.3884605407714844, 0.38845645141601565, 0.38830081176757814, 0.3884513244628906, 0.3889899597167969, 0.3887943725585937, 0.3889407958984375, 0.8262952880859376, 0.38884658813476564, 0.38921319580078123, 0.388748291015625, 0.3887923278808594, 0.38922137451171873, 0.3888977966308594, 0.3885906066894531, 0.3892326354980469, 0.38878106689453124, 0.38923776245117186, 0.38898483276367185, 0.3888721923828125, 0.38877694702148435, 0.38888754272460935, 0.3888670654296875, 0.38904830932617185, 0.3887943725585937, 0.38909030151367185, 0.38900018310546874, 0.3890636901855469, 0.3890831298828125, 0.3892695007324219, 0.38899404907226565, 0.39072256469726563, 0.3887984619140625, 0.3887677307128906, 0.3885977478027344, 0.38868377685546873, 0.38868069458007815, 0.38871551513671876, 0.3887615966796875, 0.38870834350585937, 0.38858139038085937, 0.388885498046875, 0.3888230285644531, 0.3896135559082031, 0.39009588623046876, 0.3886540832519531, 0.38893875122070315, 0.3886039123535156, 0.3890165710449219, 0.3886899108886719, 0.38870529174804686, 0.38863665771484374, 0.3886868591308594, 0.38889166259765623, 0.3890595703125, 0.38893771362304685, 0.3889479675292969, 0.388917236328125, 0.38900222778320315, 0.38909951782226565, 0.38838067626953127, 0.388495361328125, 0.3884666748046875, 0.389074951171875, 0.38862539672851565, 0.38855474853515626, 0.3885578308105469, 0.388843505859375, 0.389080078125, 0.3892039794921875, 0.8267151489257812, 0.3888455810546875, 0.389074951171875, 0.38913229370117186, 0.38876263427734375, 0.3887595520019531, 0.3888875732421875, 0.38918243408203124, 0.3892049865722656, 0.38914047241210936, 0.3887319030761719, 0.38899713134765623, 0.38865509033203127, 0.38901556396484377, 0.3886766052246094, 0.38860800170898435, 0.388864013671875, 0.38877593994140625, 0.3888302001953125, 0.388790283203125, 0.3885578308105469, 0.38857318115234374, 0.3885926513671875, 0.3890616455078125, 0.3892623291015625, 0.3890933837890625, 0.38865817260742186, 0.3889070129394531, 0.38874929809570313, 0.38892340087890626, 0.3887185974121094, 0.3888066711425781, 0.38876568603515627, 0.3885875244140625, 0.38916607666015623, 0.38873394775390624, 0.3885383605957031, 0.38869403076171877, 0.38857217407226563, 0.38951425170898435, 0.38901556396484377, 0.38918759155273436, 0.3889551391601562, 0.3887442016601563, 0.3890032653808594, 0.38889675903320314, 0.3889725341796875, 0.3889837951660156, 0.38876568603515627, 0.38967501831054685, 0.3893606262207031, 0.38946200561523436, 0.3886407775878906, 0.38881381225585937, 0.38867251586914064, 0.38852505493164063, 0.38857217407226563, 0.38876568603515627, 0.38846771240234373, 0.38870529174804686, 0.3890585632324219, 0.38904730224609374, 0.38865716552734375, 0.8268502807617187, 0.3887595520019531, 0.3887820739746094, 0.38875750732421877, 0.3886417846679687, 0.38880767822265627, 0.38890087890625, 0.38849432373046877, 0.38883224487304685, 0.3887718505859375, 0.38920294189453125, 0.3890390930175781, 0.3896995849609375, 0.38909747314453125, 0.3887442016601563, 0.388569091796875, 0.3890595703125, 0.3884718017578125, 0.38869195556640623, 0.38880563354492187, 0.38886605834960936, 0.38868377685546873, 0.3896105041503906, 0.38871450805664065, 0.3889029235839844, 0.38878311157226564, 0.3886878662109375, 0.38854757690429687, 0.38863873291015627, 0.3887370300292969, 0.38861004638671875, 0.38876263427734375, 0.3888015441894531, 0.3884984436035156, 0.38875546264648436, 0.38851174926757814, 0.388959228515625, 0.38882406616210935, 0.3891138610839844, 0.38870733642578126, 0.38867251586914064, 0.38911282348632814, 0.38889166259765623, 0.3884820556640625, 0.38871450805664065, 0.3886540832519531, 0.388600830078125, 0.389032958984375, 0.389180419921875, 0.38953677368164064, 0.38902886962890626, 0.3890667419433594, 0.389001220703125, 0.3887820739746094, 0.38870016479492187, 0.3886346130371094, 0.3887544250488281, 0.38863873291015627, 0.38883328247070315, 0.3885629577636719, 0.3887093811035156, 0.38889675903320314, 0.38913946533203125, 0.8280023193359375, 0.38870834350585937, 0.38864794921875, 0.3887124328613281, 0.388748291015625, 0.3888793640136719, 0.3885383605957031, 0.38862744140625, 0.38867352294921875, 0.3893258361816406, 0.3888793640136719, 0.38871551513671876, 0.38890802001953123, 0.38877490234375, 0.38909747314453125, 0.3890729064941406, 0.3887790222167969, 0.38876876831054685, 0.3889479675292969, 0.38905447387695313, 0.388927490234375, 0.38944769287109376, 0.3890114440917969, 0.3895459899902344, 0.3894302673339844, 0.3890831298828125, 0.38988287353515627, 0.388738037109375, 0.3887667236328125, 0.38885888671875, 0.3889950866699219, 0.3887370300292969, 0.3887564697265625, 0.38875546264648436, 0.3886991271972656, 0.389185546875, 0.38876058959960935, 0.38907086181640627, 0.38890188598632813, 0.3888619384765625, 0.38933401489257813, 0.388853759765625, 0.38896026611328127, 0.38860595703125, 0.38862744140625, 0.38865817260742186, 0.388463623046875, 0.3885875244140625, 0.3883212890625, 0.3891998596191406, 0.3895500793457031, 0.3895326843261719, 0.389032958984375, 0.3888066711425781, 0.388822021484375, 0.388811767578125, 0.38866329956054685, 0.38884146118164065, 0.38888958740234375, 0.38867147827148435, 0.38877490234375, 0.3891199951171875, 0.38873599243164064, 0.8254505004882813, 0.38886605834960936, 0.3887943725585937, 0.3888025512695312, 0.38903604125976565, 0.38877694702148435, 0.38867864990234374, 0.3889029235839844, 0.3888517150878906, 0.38892645263671877, 0.38902374267578127, 0.38975079345703123, 0.389823486328125, 0.389465087890625, 0.3888455810546875, 0.3890862121582031, 0.38899200439453124, 0.388611083984375, 0.3885342712402344, 0.3887175598144531, 0.38892340087890626, 0.38893875122070315, 0.3891138610839844, 0.38856704711914064, 0.3884707946777344, 0.3899740295410156, 0.388389892578125, 0.388780029296875, 0.3886213073730469, 0.3885793151855469, 0.3890093994140625, 0.38884454345703123, 0.3886991271972656, 0.388748291015625, 0.3889151916503906, 0.38854757690429687, 0.3891988525390625, 0.3890667419433594, 0.388822021484375, 0.38893670654296875, 0.3886960754394531, 0.38903399658203125, 0.38867251586914064, 0.38844415283203126, 0.3887595520019531, 0.3886796875, 0.3885424499511719, 0.3885875244140625, 0.3885660095214844, 0.388421630859375, 0.38844927978515625, 0.3885557861328125, 0.3888015441894531, 0.3884236755371094, 0.38832846069335936, 0.388347900390625, 0.388389892578125, 0.38839910888671875, 0.38838580322265626, 0.38835302734375, 0.3885363159179688, 0.38859982299804685, 0.38915277099609374, 0.8270069580078125, 0.388790283203125, 0.38852301025390623, 0.38860800170898435, 0.38870529174804686, 0.3886397399902344, 0.3888097229003906, 0.38888140869140625, 0.38900531005859373, 0.3891435546875, 0.38884454345703123, 0.38879129028320314, 0.38904934692382814, 0.38878924560546874, 0.38913946533203125, 0.38866329956054685, 0.3886970825195312, 0.3890196533203125, 0.38881076049804686, 0.38890087890625, 0.3889612731933594, 0.38866021728515626, 0.3888097229003906, 0.38903604125976565, 0.38920907592773435, 0.389138427734375, 0.38885888671875, 0.38871142578125, 0.38858648681640623, 0.3885793151855469, 0.38836428833007813, 0.38872164916992186, 0.3887953796386719, 0.3888721923828125, 0.389064697265625, 0.3893247985839844, 0.3885936584472656, 0.38854757690429687, 0.3885936584472656, 0.3884789733886719, 0.38885479736328127, 0.38876058959960935, 0.38854656982421876, 0.3885404052734375, 0.38906982421875, 0.38860287475585936, 0.38849945068359376, 0.38863665771484374, 0.38878106689453124, 0.3889336242675781, 0.38914559936523435, 0.38887322998046875, 0.3893893127441406, 0.3887462463378906, 0.38847796630859377, 0.38849127197265626, 0.38849740600585936, 0.3886684265136719, 0.3886510009765625, 0.38863565063476563, 0.38843597412109376, 0.3889039306640625, 0.38858853149414063, 0.8253368530273437, 0.38870220947265627, 0.38867864990234374, 0.3885823974609375, 0.3886929931640625, 0.3884881896972656, 0.38854656982421876, 0.38842059326171874, 0.38854757690429687, 0.38858648681640623, 0.388569091796875, 0.3887452087402344, 0.3891435546875, 0.38859982299804685, 0.3888302001953125, 0.38854144287109377, 0.3885936584472656, 0.38863052368164064, 0.38876980590820315, 0.3885404052734375, 0.38905447387695313, 0.38915277099609374, 0.38901556396484377, 0.38856704711914064, 0.3885404052734375, 0.38846157836914064, 0.38859469604492186, 0.38851071166992185, 0.38895001220703124, 0.38856805419921875, 0.38871450805664065, 0.38837246704101563, 0.3885987854003906, 0.38849127197265626, 0.38861618041992185, 0.38849639892578125, 0.38856192016601565, 0.38879641723632813, 0.38922341918945313, 0.38905752563476564, 0.38980096435546874, 0.3889090576171875, 0.3889070129394531, 0.3885916137695313, 0.38894183349609374, 0.3886796875, 0.3886407775878906, 0.3885383605957031, 0.38921829223632814, 0.38887322998046875, 0.38906265258789063, 0.3887585144042969, 0.38897869873046875, 0.38905548095703124, 0.38911077880859374, 0.3886458740234375, 0.3885455322265625, 0.3886878662109375, 0.3890585632324219, 0.38870220947265627, 0.38884658813476564, 0.38864486694335937, 0.3889561462402344, 0.82583349609375, 0.3893094482421875, 0.38874725341796873, 0.38864697265625, 0.3890738525390625, 0.3888230285644531, 0.3885772705078125, 0.3886264343261719, 0.38845541381835935, 0.38850457763671875, 0.38880767822265627, 0.3886039123535156, 0.3885486145019531, 0.388485107421875, 0.3883673706054688, 0.3887646789550781, 0.3885353088378906, 0.388632568359375, 0.38862539672851565, 0.3885936584472656, 0.38859674072265626, 0.3886069641113281, 0.388485107421875, 0.38892440795898436, 0.38891622924804686, 0.38909133911132815, 0.388453369140625, 0.38876058959960935, 0.3886592102050781, 0.3885567932128906, 0.3885189208984375, 0.3885557861328125, 0.388716552734375, 0.3882700805664063, 0.38850457763671875, 0.3887790222167969, 0.38870733642578126, 0.3885281372070313, 0.38864382934570313, 0.3885977478027344, 0.38863565063476563, 0.3885660095214844, 0.38846463012695315, 0.38862234497070314, 0.3884236755371094, 0.3884984436035156, 0.388347900390625, 0.38830181884765624, 0.38828134155273436, 0.3885066223144531, 0.38897152709960936, 0.38908721923828127, 0.3887431640625, 0.3887175598144531, 0.3884800109863281, 0.38863665771484374, 0.39151205444335935, 0.3889510498046875, 0.38859674072265626, 0.38852301025390623, 0.38847796630859377, 0.3888824462890625, 0.38877593994140625, 0.824754150390625, 0.38834994506835935, 0.38857217407226563, 0.38853323364257814, 0.3888844909667969, 0.3888005065917969, 0.3891640319824219, 0.3885987854003906, 0.3888486328125, 0.38872781372070314, 0.38896435546875, 0.3892623291015625, 0.3890882568359375, 0.3891937255859375, 0.38884658813476564, 0.3889827880859375, 0.38853323364257814, 0.3885260925292969, 0.38840933227539065, 0.38836428833007813, 0.388558837890625, 0.38827621459960937, 0.3885209655761719, 0.3886376953125, 0.38875341796875, 0.3887790222167969, 0.38885888671875, 0.3887298583984375, 0.38882406616210935, 0.38858035278320313, 0.388927490234375, 0.3888988037109375, 0.38889166259765623, 0.38841650390625, 0.3884298095703125, 0.3886039123535156, 0.3884247131347656, 0.3889837951660156, 0.38898175048828126, 0.3884728393554687, 0.3891671142578125, 0.3885987854003906, 0.3886120910644531, 0.388411376953125, 0.38858139038085937, 0.3884390258789063, 0.3885076599121094, 0.388558837890625, 0.388738037109375, 0.38887322998046875, 0.38831716918945314, 0.3885014953613281, 0.3886868591308594, 0.38870220947265627, 0.38861721801757815, 0.38846258544921874, 0.3886673889160156, 0.3884789733886719, 0.38866943359375, 0.388358154296875, 0.3887667236328125, 0.38856500244140624, 0.388558837890625]",tokens/s,2.5313503159658475,, @@ -2010,7 +2010,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e48d-429c38ef7d866605540f2936;c0438915-b7b1-4ac9-b8d5-43ed841fa18c) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722cd9-226302d73ab8e5154364e9bc;c1fbae43-ce1f-465f-b77b-364eabdc6e26) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2068,7 +2068,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmptox0y53l/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp2zc01gxl/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v1-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,facebook/opt-350m,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,1,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1279.791104,1019.74016,0.0,373.293056,277.410816,s,10,0.3806483840942383,0.03806483840942383,0.0010374051807092283,0.03839825630187988,0.038631001281738284,0.03909522857666015,0.03946661041259766,"[0.03955945587158203, 0.03852783966064453, 0.03837411117553711, 0.038422401428222654, 0.03847126388549805, 0.038292991638183595, 0.03834467315673828, 0.03843852615356445, 0.036084991455078125, 0.03613212966918945]",tokens/s,6725.366787229584,kWh,4.3617786259962035e-07,2.3900503699099743e-07,1.1732389459702955e-06,1.8484218455609134e-06,tokens/kWh,138496523.73174343,MB,1280.086016,1019.74016,0.0,373.293056,323.047424,s,10,23.712255126953128,2.371225512695313,0.003882883813537052,2.3718930664062503,2.376249291992188,2.3762570922851562,2.3762633325195313,"[2.362728515625, 2.37624755859375, 2.373291748046875, 2.367230224609375, 2.371970947265625, 2.371815185546875, 2.376264892578125, 2.37161865234375, 2.37223681640625, 2.3688505859375]",tokens/s,26.56854004931377,kWh,2.7777078536237154e-05,1.5222735153430559e-05,6.781520513363374e-05,0.00011081501882330147,tokens/kWh,568514.9961527847,,s,629,24.01549103546141,0.03818043089898478,0.004509392475924156,0.03761459350585938,0.03786444854736328,0.03806208114624023,0.07505829895019543,"[0.03638784027099609, 0.03659775924682617, 0.036880382537841795, 0.03663974380493164, 0.03670937728881836, 0.03717529678344727, 0.036736000061035154, 0.03655987167358398, 0.03629260635375976, 0.03701657485961914, 0.03663974380493164, 0.0363612174987793, 0.03665100860595703, 0.038844417572021485, 0.03839078521728516, 0.038013950347900394, 0.03768627166748047, 0.03772825622558594, 0.037766143798828124, 0.03749990463256836, 0.037585918426513674, 0.03749683380126953, 0.03755929565429687, 0.03760537719726562, 0.037743614196777346, 0.03752140808105469, 0.0375551986694336, 0.03762995147705078, 0.037495807647705076, 0.03751935958862305, 0.03787571334838867, 0.03770880126953125, 0.0376360969543457, 0.03753472137451172, 0.037599231719970705, 0.03759718322753906, 0.03856595230102539, 0.03801900863647461, 0.037601280212402347, 0.03762073516845703, 0.0376545295715332, 0.03752243041992188, 0.03753881454467774, 0.03760435104370117, 0.03775795364379883, 0.037550079345703126, 0.03786649703979492, 0.03831398391723633, 0.03769036865234375, 0.037601280212402347, 0.03755929565429687, 0.037498878479003905, 0.03746303939819336, 0.03752345657348633, 0.03794432067871094, 0.037713920593261716, 0.03760435104370117, 0.037495807647705076, 0.037591041564941405, 0.03748454284667969, 0.03763916778564453, 0.03788083267211914, 0.07681126403808594, 0.03763711929321289, 0.03831808090209961, 0.03748966217041016, 0.03763507080078125, 0.03758489608764649, 0.03759308624267578, 0.037579776763916016, 0.03772419357299805, 0.03755926513671875, 0.03763097763061524, 0.03747020721435547, 0.03756748962402344, 0.03761663818359375, 0.037901313781738284, 0.03767398452758789, 0.03793817520141601, 0.03772415924072266, 0.037601280212402347, 0.03751321411132812, 0.03760332870483398, 0.03770470428466797, 0.03760435104370117, 0.03767705535888672, 0.03810611343383789, 0.0376995849609375, 0.03758694458007812, 0.03785728073120117, 0.03757878494262695, 0.03765244674682617, 0.03786956787109375, 0.03756032180786133, 0.037838848114013675, 0.037773311614990236, 0.037525505065917966, 0.037884929656982425, 0.03775692749023438, 0.03761151885986328, 0.0378869743347168, 0.03772518539428711, 0.037797889709472655, 0.03774566268920899, 0.037759998321533206, 0.03785113525390625, 0.038040576934814455, 0.03770470428466797, 0.0377077751159668, 0.03755929565429687, 0.037599231719970705, 0.037661697387695314, 0.037553150177001955, 0.037823486328125, 0.03759820938110352, 0.03769651031494141, 0.037648384094238284, 0.03772313690185547, 0.03791667175292969, 0.03774259185791016, 0.03760435104370117, 0.03763711929321289, 0.03762483215332031, 0.03786444854736328, 0.03789516830444336, 0.07634636688232421, 0.03765350341796875, 0.03746713638305664, 0.03770470428466797, 0.03758796691894531, 0.03751116943359375, 0.03785318374633789, 0.037889022827148434, 0.03752345657348633, 0.0374774398803711, 0.037649345397949216, 0.03765760040283203, 0.037743614196777346, 0.0377149429321289, 0.03764329528808594, 0.03746198272705078, 0.037561344146728515, 0.037763072967529294, 0.03768832015991211, 0.03759823989868164, 0.037550048828125, 0.03745382308959961, 0.03772825622558594, 0.03768012619018555, 0.03773952102661133, 0.03763097763061524, 0.03759820938110352, 0.03765657424926758, 0.0381921272277832, 0.037905406951904294, 0.03770675277709961, 0.03752345657348633, 0.03779072189331055, 0.0376627197265625, 0.03760844802856445, 0.03748761749267578, 0.03757670211791992, 0.0375654411315918, 0.03750713729858399, 0.03755820846557617, 0.0375551986694336, 0.03763302230834961, 0.037645313262939455, 0.03746099090576172, 0.03763507080078125, 0.0376258544921875, 0.03769651031494141, 0.0375203857421875, 0.037610496520996094, 0.037612545013427735, 0.03768627166748047, 0.03968307113647461, 0.03810201644897461, 0.03744870376586914, 0.03748761749267578, 0.03760844802856445, 0.037705726623535156, 0.0375654411315918, 0.037533695220947266, 0.037482494354248046, 0.037713920593261716, 0.037579776763916016, 0.037495807647705076, 0.07634534454345702, 0.03747840118408203, 0.037424129486083986, 0.03769753646850586, 0.03747020721435547, 0.03760639953613281, 0.03746815872192383, 0.03759308624267578, 0.03759001541137695, 0.037582847595214845, 0.037572608947753904, 0.037525505065917966, 0.03752959823608398, 0.03775590515136719, 0.03758899307250976, 0.037459968566894535, 0.0377446403503418, 0.037495807647705076, 0.03755929565429687, 0.037438465118408204, 0.03756851196289063, 0.037602302551269534, 0.037515262603759765, 0.03751321411132812, 0.03760639953613281, 0.037541889190673826, 0.03759206390380859, 0.03758694458007812, 0.03766681671142578, 0.037548030853271484, 0.037512191772460936, 0.037582847595214845, 0.037493759155273435, 0.037449726104736326, 0.03746815872192383, 0.03757056045532227, 0.037823486328125, 0.037482494354248046, 0.037579776763916016, 0.03767091369628906, 0.03766886520385742, 0.03751424026489258, 0.03748147201538086, 0.037495807647705076, 0.03765555191040039, 0.03773132705688476, 0.03749273681640625, 0.03760639953613281, 0.03768217468261719, 0.037533695220947266, 0.037394432067871096, 0.03765350341796875, 0.03762278366088867, 0.03748966217041016, 0.03768729782104492, 0.0376514892578125, 0.03766985702514648, 0.03745280075073242, 0.037541889190673826, 0.03762073516845703, 0.03751731109619141, 0.037572608947753904, 0.03764223861694336, 0.07663206481933593, 0.037482494354248046, 0.03758694458007812, 0.03754598236083984, 0.03749785614013672, 0.03743641662597656, 0.03762790298461914, 0.03742105484008789, 0.03755929565429687, 0.037628929138183595, 0.03765350341796875, 0.03744768142700195, 0.03746918487548828, 0.03767500686645508, 0.03757670211791992, 0.03751222229003906, 0.03749884796142578, 0.03762995147705078, 0.03771596908569336, 0.03759718322753906, 0.03778252792358398, 0.03771596908569336, 0.03758182525634766, 0.03766886520385742, 0.0377262077331543, 0.03749683380126953, 0.0374917106628418, 0.037571582794189456, 0.03759513473510742, 0.03750912094116211, 0.037544960021972655, 0.0376627197265625, 0.03753574371337891, 0.037466110229492186, 0.03755212783813477, 0.03748044967651367, 0.03741798400878906, 0.03748761749267578, 0.03759414291381836, 0.03739030456542969, 0.03777740859985351, 0.03743129730224609, 0.037466110229492186, 0.037466110229492186, 0.03757056045532227, 0.03745177459716797, 0.03782451248168945, 0.03930624008178711, 0.03780198287963867, 0.037553150177001955, 0.037572608947753904, 0.037628929138183595, 0.03765350341796875, 0.03766579055786133, 0.03758796691894531, 0.03759718322753906, 0.03784396743774414, 0.038711296081542966, 0.03820236968994141, 0.037833728790283204, 0.03772931289672852, 0.03775075149536133, 0.03758796691894531, 0.07656038665771485, 0.03744153594970703, 0.037580799102783204, 0.03758489608764649, 0.03763507080078125, 0.03759308624267578, 0.03766681671142578, 0.0375470085144043, 0.0374835205078125, 0.03772313690185547, 0.037703678131103514, 0.037541889190673826, 0.0375551986694336, 0.037652511596679684, 0.037497825622558593, 0.03801190567016602, 0.037644287109375, 0.03753881454467774, 0.03759718322753906, 0.03752345657348633, 0.037541889190673826, 0.03757363128662109, 0.0374835205078125, 0.03762176132202148, 0.03808870315551758, 0.03783782577514649, 0.03767705535888672, 0.03754086303710937, 0.03700121688842774, 0.03764223861694336, 0.03752959823608398, 0.037765121459960936, 0.03765862274169922, 0.03768012619018555, 0.03785116958618164, 0.03752854537963867, 0.03746201705932617, 0.037591041564941405, 0.03764022445678711, 0.03770057678222656, 0.03760435104370117, 0.037569534301757815, 0.03757875061035156, 0.037550079345703126, 0.03768320083618164, 0.03784806442260742, 0.037806079864501956, 0.037579776763916016, 0.03758796691894531, 0.03765248107910156, 0.03765862274169922, 0.03774566268920899, 0.03764121627807617, 0.03772313690185547, 0.03784294509887695, 0.037817344665527344, 0.03783987045288086, 0.03761971282958984, 0.03763814544677734, 0.03765862274169922, 0.0377077751159668, 0.03765862274169922, 0.03762995147705078, 0.07627571105957032, 0.03767603302001953, 0.03746713638305664, 0.03762278366088867, 0.03759001541137695, 0.037689342498779296, 0.03778559875488281, 0.03764019012451172, 0.03788390350341797, 0.03751424026489258, 0.03749273681640625, 0.03770880126953125, 0.03754291152954101, 0.037703678131103514, 0.03769036865234375, 0.037574657440185545, 0.03785420989990235, 0.03765657424926758, 0.03755929565429687, 0.037599231719970705, 0.03741593551635742, 0.03765555191040039, 0.0375838737487793, 0.038059009552001956, 0.03752243041992188, 0.03781324768066406, 0.03762790298461914, 0.03782761764526367, 0.03764118576049805, 0.03759513473510742, 0.03786444854736328, 0.037617664337158206, 0.03751731109619141, 0.037566463470458986, 0.03778252792358398, 0.037596160888671876, 0.03751321411132812, 0.03806412887573242, 0.03936972808837891, 0.03786342239379883, 0.03775897598266602, 0.037601280212402347, 0.03762483215332031, 0.03776921463012695, 0.03782860946655273, 0.037449726104736326, 0.037705726623535156, 0.0375654411315918, 0.03755929565429687, 0.03753267288208008, 0.03763302230834961, 0.03753881454467774, 0.037580799102783204, 0.03768832015991211, 0.037574657440185545, 0.037531646728515625, 0.03764940643310547, 0.03754393768310547, 0.03761151885986328, 0.037515262603759765, 0.03761459350585938, 0.03788288116455078, 0.03977830505371094, 0.07646412658691407, 0.03738828659057617, 0.03792588806152344, 0.037748737335205076, 0.03746713638305664, 0.037897216796875, 0.037553150177001955, 0.03762176132202148, 0.037343231201171875, 0.037651454925537106, 0.037405696868896485, 0.0374835205078125, 0.03782758331298828, 0.03766681671142578, 0.03759001541137695, 0.0375838737487793, 0.0378081283569336, 0.03773235321044922, 0.03765760040283203, 0.03766886520385742, 0.03765964889526367, 0.03771187210083008, 0.03750707244873047, 0.03753472137451172, 0.037531646728515625, 0.03755110549926758, 0.03769343948364258, 0.03748659133911133, 0.037596160888671876, 0.03750707244873047, 0.03753881454467774, 0.037820415496826174, 0.03772415924072266, 0.03755929565429687, 0.037553150177001955, 0.03769036865234375, 0.03775897598266602, 0.037594112396240234, 0.03757875061035156, 0.03759513473510742, 0.03752755355834961, 0.03780505752563477, 0.03753267288208008, 0.03759820938110352, 0.03762790298461914, 0.03850447845458985, 0.03777225494384766, 0.03763916778564453, 0.03750297546386719, 0.037561344146728515, 0.037645313262939455, 0.03790848159790039, 0.037582847595214845, 0.037553150177001955, 0.037664768218994144, 0.03753574371337891, 0.03770675277709961, 0.03749075317382813, 0.03756332778930664, 0.03781324768066406, 0.03759718322753906, 0.037773311614990236, 0.037571582794189456, 0.07186124420166015, 0.037190654754638675, 0.03747225570678711, 0.03754908752441406, 0.037579742431640625, 0.03760639953613281, 0.03770982360839844, 0.03739136123657227, 0.03773440170288086, 0.0376360969543457, 0.03769241714477539, 0.03769241714477539, 0.03752345657348633, 0.03768627166748047, 0.03762688064575195, 0.03752140808105469, 0.037733375549316404, 0.03770675277709961, 0.037754878997802735, 0.03753472137451172, 0.03745792007446289, 0.03804876708984375, 0.03773747253417969, 0.037572608947753904, 0.03754291152954101, 0.03757056045532227, 0.03814297485351562, 0.03765350341796875, 0.03766988754272461, 0.03759001541137695, 0.03777536010742188, 0.03753472137451172, 0.037574657440185545, 0.03759718322753906, 0.03750707244873047, 0.03766886520385742, 0.03754393768310547, 0.03745280075073242, 0.03773952102661133, 0.03783168029785156, 0.03754086303710937, 0.03762688064575195, 0.037483551025390624, 0.03754902267456055, 0.03769036865234375, 0.03759718322753906, 0.037456897735595705, 0.03748761749267578, 0.038013950347900394, 0.037768192291259765, 0.03762073516845703, 0.03767705535888672, 0.037928958892822266, 0.03753779220581055, 0.03802828979492188, 0.037585918426513674, 0.039975936889648435, 0.03809791946411133, 0.03769343948364258, 0.037564414978027344, 0.03762278366088867, 0.03766988754272461, 0.03770470428466797, 0.07192781066894531, 0.03523993682861328, 0.03559731292724609, 0.038373374938964845, 0.03784908676147461, 0.03781017684936523, 0.03761663818359375, 0.037765121459960936, 0.037564414978027344, 0.037615615844726565, 0.03755417633056641, 0.038040576934814455, 0.03775692749023438, 0.03775692749023438, 0.037582847595214845, 0.037716991424560545, 0.03759308624267578, 0.03775283050537109, 0.03753062438964844, 0.03760435104370117, 0.03752447891235351, 0.037569534301757815, 0.03765964889526367, 0.03779891204833984, 0.03758182525634766, 0.037698558807373043, 0.037585918426513674, 0.03753062438964844, 0.03767398452758789, 0.037664768218994144, 0.0376627197265625, 0.037787647247314454, 0.03765657424926758, 0.037833728790283204, 0.037781505584716796, 0.037684223175048825, 0.037719039916992186, 0.03794944000244141, 0.037571582794189456, 0.03768627166748047, 0.037579776763916016, 0.037574657440185545, 0.03738521575927734, 0.03818188858032227, 0.0376627197265625, 0.03749990463256836, 0.037795841217041014, 0.03761151885986328, 0.037803009033203126, 0.0377077751159668, 0.03760435104370117, 0.03778867340087891, 0.037615615844726565, 0.03751731109619141, 0.03779379272460937, 0.03823820877075195, 0.03774054336547852, 0.03762688064575195, 0.03750400161743164, 0.037553150177001955, 0.037956607818603515, 0.03771289443969727, 0.037722110748291016]",tokens/s,26.191427819286087,, @@ -2232,7 +2232,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e430-5468c04e726e2c74679a4228;0405280a-e9fc-4210-aa33-38635889f343) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722c88-3fe3f16c5270d514050d2331;bc3b3066-b01c-4dbe-a1de-21bafb5b170f) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2302,7 +2302,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e8d5-70e89f185727510c73738967;3cc09ca1-ae96-460e-996e-5b8cf90dc326) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667230ce-38f66e682c37960645e6381b;23e5bd09-6a20-40ba-8071-a4bd99286653) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2372,7 +2372,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e872-08abaf584811ad3f0149a204;89237cb0-6278-4e68-91eb-733c953f957a) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6672307e-28f791b703bf3c57108fa08a;b11b6607-d8fc-4679-a725-d163da04e8fd) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2472,7 +2472,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e730-459ae65c7fba894471a43ff1;95fec30d-4670-4064-b091-0e2ac3e59557) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722f54-733acb265999f01678f8479f;7d674b4c-92a1-4750-b4b4-c4e08f0ffa23) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2571,7 +2571,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e7a9-7c43db445544a3ef296e8968;dd9cad12-58b5-4630-9bfb-c5f87fc1223a) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722fcf-50857cce625af523504274ad;18f3e479-13a2-418b-a8fc-4bb5c5ec8ceb) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -2839,7 +2839,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d801-77baef2c69b0a7b721ddb518;ef0f1c61-49db-4c8f-ae37-2cd8992f2ac9) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6672298e-0fe0ee937f11e7c22bbb0d64;2977ea84-8509-4728-9985-9d31db5e8705) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -3044,7 +3044,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e759-35c37dd32d6067913dbfad81;58e5bb18-baf8-45da-b42b-464d971006cb) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722f7b-6bb9f606518aee5d11f0c6b8;9a014b98-eec1-43d9-a65a-9ecefdf6c8d8) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -3104,7 +3104,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d7d8-41f1c07828feda1e49441ab2;422d2d3f-607d-4023-89cc-41c914d48a53) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66722966-49e260f97d49e976490e5688;c1a63352-6f04-4db5-b95e-1ccd01de7655) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -3629,7 +3629,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d63f-70f76ecd412f5301068afe41;209659a7-e8dc-4734-ba46-8a3ce588de12) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667227dd-4dd92ad117f93d9e4909e664;4ae4e3a0-58f5-4dca-b628-947f4f7dd24d) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -3714,7 +3714,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e6e0-5bee696633c35abb10fa0a95;9727e1cb-6a9e-485f-a0c7-d7f9631049d8) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722f04-7a7605836b93d4dd388583cb;20c0132f-1faa-46d9-bcd6-2c6dd5e2da05) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -3858,7 +3858,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e617-0b5658260d07187538508fdb;6d985839-5386-4cfb-8971-763cde79472e) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722e3b-376f595f21ff9ed860bb2758;d2513add-ab4d-495b-b118-eb83fed8328b) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4030,7 +4030,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e82e-25aa29b72542f13b579f7793;091c0572-a6d5-4097-8654-2f98d056d8d5) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66723047-72a84794428d3f976ec58b19;e7c51480-0605-4a76-8e77-5f63ace4843e) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4129,7 +4129,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e5ee-1f6a30f35289df7832a44b68;a953ad48-e602-449a-90b7-d72db92354f0) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722e13-196212c24eebd55e5e4d583a;3a49ed87-2ea9-4402-b883-294006aee1b2) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4228,7 +4228,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e68f-3e23550579095d4754821fb8;40673687-a4b7-48ec-a9ff-0f499f09ba4d) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722eb4-77e929bb233b50244275e1de;3e7faf11-162c-4d3a-a10e-95f491314815) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4319,7 +4319,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e9d5-712824e953f911ef7e5c58ca;6968df34-8595-4dc4-ba9c-1dfac596ec7d) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667231c2-5161092633cf2e706000ace1;d8151e84-01f9-43b1-8864-eec98c91e58a) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -4404,7 +4404,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e6b7-14b3df764cc7dd967d2456d9;32b0290b-8046-436d-8293-0a4440113c59) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722edc-395375305df2a31c3aa6a9d4;9c888b84-b4d7-482d-abac-65e1c522ebbe) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4494,7 +4494,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d7b1-12f8e71a475ed55d697c7d62;c9fa4803-9eb4-4367-ab6c-5f9e46b54d1c) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6672293e-0546debf2c7adb665e2176a6;0b62ee5c-2e34-411f-8e83-be180cdce408) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -4677,7 +4677,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e667-77e7adae7899b3d4220679d8;296772ba-550f-4fa1-844b-f2e67a133a88) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722e8b-73ddec451897e3cb4f6f0597;6265c566-e1bc-45e6-b580-3889f5f016ae) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4806,7 +4806,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670d829-3b52238a3e27ab8f400fd2b6;f9ed387b-3879-42f2-82a4-12c5f57f0364) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667229b8-03cbdaec472d898d6dfa4361;3b8b0c94-6cbc-40f1-8577-499f9a7c2fc8) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -4935,7 +4935,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e7d1-1e05ac6950a6170949aaf0e1;2c62c82c-8a82-4207-9a9e-83446e50f4f1) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722ff8-4a72a14e132898af2d8e01b6;4333789d-c5e8-4cc4-9782-bc0631586528) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5035,7 +5035,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e47f-28f7df0f51bf57b4324c6ede;8e34a220-ee5f-4845-942d-2e3841445641) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722ccc-6cec8ba56ced6b30678c8269;f2b002b1-ba83-4e40-a576-6568abe4c7dc) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5384,7 +5384,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e422-6368faee2a69faa659052e39;78261a00-53e7-49a8-b092-66a4831706ec) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722c7a-3248681d0669d81e0ddb0821;deb56ce6-9b96-4501-8492-98b922f1f4e2) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5454,7 +5454,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e8c6-5ed177994f68f8c37f435b66;cb751be2-b935-4c13-9733-9fc5e8993516) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667230c1-320803f61f1caae5012c8030;dea81254-b97d-4541-9106-1a01aa4d2724) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5524,7 +5524,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e85c-6f55982815d9ab5656deb828;895dd110-0f05-4ef6-bd82-617cf631a08a) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66723070-39f9891e245afed77c3d79f2;379ff680-9011-4823-9645-82846fccdde8) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5652,7 +5652,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e744-624d3e654ad9a6230b8b6201;4d60829d-6e37-477a-b6d7-77cd884d3b8d) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722f68-3bc34f2e4104ef3f2cfbe31f;1c79a818-c079-462f-81f1-5c926a1d650a) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5723,7 +5723,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e7bd-1082501e2c7375471e486069;c606cf57-e9f0-47ec-9d39-04b4ab229bb3) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722fe3-5a8af8cb7780c4aa42387dd6;425c8f03-b445-4a3c-897f-735336de283b) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -5935,7 +5935,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d815-0df933c6614bd43b0647cdd5;799aac47-7107-45e3-beb8-543600571bda) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667229a3-74bb30c0738b50fa32f1f12b;afc2290f-1a2e-4986-a858-4287952756f9) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -6024,7 +6024,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e76d-3ace38d106192f36169a22c6;2ba6e9be-4d89-4762-8af5-c929a5a87c9f) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722f8f-58107c6a25f0ab6b29943259;82978151-5b62-475b-b85a-1225db7a702a) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -6084,7 +6084,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d7ec-2ea8cc481e880d08729bd4c3;3f919044-c952-46fd-839e-69a0c748d790) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6672297a-1e7ca7a4570287391d9e1b25;4bbbf92c-6f3d-4d84-9b3b-990f12fd1e03) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -6316,7 +6316,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpyxpdkjom/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: CodeGenForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp7cz98zvh/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen1.5-MoE-A2.7B,Qwen/Qwen1.5-MoE-A2.7B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -6413,7 +6413,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d653-278913bc5fcf8c253734c9b8;39569a41-e614-4365-ae7a-b3dbe2f536a7) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667227f1-6710128970c942fa1b173043;7a832433-45e4-4e74-ba1b-c440009dfbd6) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -6498,7 +6498,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e6f4-15ce0a8d2753be140373e6f7;febfa986-d5a5-4188-8368-ccc67557dc46) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722f18-2b9cddb10ce018c8567976de;5fa0217a-c1c7-4774-be63-0c45744c988f) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -6614,7 +6614,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e62b-7bc0d91a28c2e4cc69be0a39;4d068fd4-3340-4559-a77c-82a38f1bc932) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722e4f-3469af766bc2ba9520967c4e;2b9a9198-8e43-4d68-b172-113c2203ba0e) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -6714,7 +6714,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmps3ykffd7/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmppe8m4v27/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,Qwen/Qwen-14B,Qwen/Qwen-14B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -6786,7 +6786,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e846-34f6aeb84ec5b559427a1d1c;5f02c05d-0996-46a9-bebb-8585927801bf) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6672305c-584ad75a56f9a1dc6d37d95b;8debe7a5-80c0-4233-b2f9-fead3c0a3e80) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -6843,7 +6843,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmprh10k01q/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp77g_6quk/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,M,M,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -6885,7 +6885,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e602-619eba1a4b1e950c22186d0d;d05b1b81-1ab3-436b-bb83-67e5a654ec3d) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722e26-56469f4b5479a10b511f3683;7e5f4fe6-6ebd-4871-a8c3-c3910847bde1) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -6956,7 +6956,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e6a3-29b3da8035dda9d864f74a0e;82a1f71d-ab2a-4aa2-aba8-e042570f73b9) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722ec8-5f77d83d554f61623f10736d;0cfe4740-45dd-466b-96d5-0d5dbc08d61d) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7014,7 +7014,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmplwwieoyi/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmp_fetbaq6/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,01-ai/Yi-34B,,cuda,0,42,,,,,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.0,217063f5c507ed7cc255df7e1f64c4333a0b4dfe,4.40.2,,0.30.1,,,,1.19.2,,,,0.10.0,,,main,False,False,MB,4755.23072,21498.42944,0.0,20851.982336,20236.60032,s,10,15.643266601562498,1.5643266601562498,0.001716851893423332,1.563902587890625,1.5660734985351563,1.567352362060547,1.5683754528808596,"[1.5631807861328124, 1.5647735595703125, 1.5629490966796875, 1.56462890625, 1.563903564453125, 1.5627757568359375, 1.565789306640625, 1.563901611328125, 1.5686312255859376, 1.5627327880859374]",tokens/s,163.6486844598237,kWh,1.846329511867629e-05,1.0117909425098331e-05,9.183712902519626e-05,0.00012041833356897088,tokens/kWh,2125922.1284055826,MB,4755.23072,21498.42944,0.0,20851.982336,20339.706368,s,10,879.77465625,87.97746562500001,0.0022909340685085054,87.97753515625,87.97981953125,87.980624609375,87.981268671875,"[87.974140625, 87.978578125, 87.9789609375, 87.9772265625, 87.979640625, 87.97715625, 87.97784375, 87.97353125, 87.9814296875, 87.9761484375]",tokens/s,0.716092462455496,kWh,0.0010385792660216491,0.0005692333667929961,0.005151564232359207,0.006759376865173853,tokens/kWh,9320.38577765846,,s,629,892.4578986816405,1.4188519851854382,0.185906583948783,1.396368408203125,1.3966589599609376,1.3968528564453124,2.960364462890625,"[1.39618505859375, 1.39667041015625, 1.3964708251953124, 1.396378662109375, 1.3963243408203125, 1.3962271728515625, 1.3963038330078126, 1.3964288330078125, 1.3964349365234374, 1.396369384765625, 1.3964697265625, 1.3965823974609375, 1.3962998046875, 1.3964708251953124, 1.3964349365234374, 1.3963857421875, 1.396304931640625, 1.3962802734375, 1.3960714111328125, 1.396305908203125, 1.3961553955078125, 1.3963427734375, 1.396221923828125, 1.3963756103515625, 1.39614111328125, 1.396317138671875, 1.3963294677734375, 1.3963673095703124, 1.396279296875, 1.3963018798828124, 1.3959813232421876, 1.3964195556640624, 1.396304931640625, 1.396304931640625, 1.396275146484375, 1.3961236572265625, 1.3962794189453125, 1.396401123046875, 1.39871435546875, 1.396484130859375, 1.396558837890625, 1.396505615234375, 1.3966868896484375, 1.3962772216796875, 1.3962833251953124, 1.3960928955078125, 1.396380615234375, 1.3968824462890626, 1.3964442138671875, 1.396285400390625, 1.3964237060546876, 1.396358154296875, 1.396305908203125, 1.3966787109375, 1.3965537109375, 1.3964185791015624, 1.3962916259765625, 1.3968353271484375, 1.396358154296875, 1.396221923828125, 1.39633251953125, 1.396252685546875, 2.962231201171875, 1.396673583984375, 1.3970401611328125, 1.396295654296875, 1.3965557861328124, 1.396348876953125, 1.3965035400390624, 1.3961922607421875, 1.3968536376953125, 1.39629052734375, 1.396274169921875, 1.396199462890625, 1.3962281494140625, 1.3965833740234375, 1.396538330078125, 1.3962445068359375, 1.3963079833984375, 1.3963243408203125, 1.396515869140625, 1.3962762451171875, 1.396073486328125, 1.3978162841796875, 1.396114501953125, 1.396137939453125, 1.3962198486328126, 1.3962987060546874, 1.3961595458984375, 1.3963714599609376, 1.396452392578125, 1.3961502685546876, 1.3962720947265626, 1.3963868408203124, 1.3964185791015624, 1.3962239990234375, 1.396464599609375, 1.3964482421875, 1.396427734375, 1.396380615234375, 1.396588623046875, 1.3964810791015625, 1.3967718505859374, 1.3981634521484374, 1.3962802734375, 1.396275146484375, 1.3962301025390624, 1.396253662109375, 1.396327392578125, 1.3961728515625, 1.3966571044921876, 1.396474853515625, 1.3965128173828125, 1.3963858642578124, 1.396339599609375, 1.3961318359375, 1.3965147705078125, 1.3963387451171876, 1.396506591796875, 1.3962393798828125, 1.3964605712890625, 1.3965179443359375, 1.39639599609375, 1.3985382080078126, 1.3965301513671875, 2.9603779296875, 1.396199462890625, 1.3967340087890625, 1.3962291259765625, 1.3964554443359376, 1.396252685546875, 1.3964892578125, 1.3960601806640625, 1.39671240234375, 1.3963714599609376, 1.3963509521484374, 1.3961092529296875, 1.39653125, 1.3965814208984375, 1.3964400634765626, 1.3963172607421874, 1.39658349609375, 1.3963939208984375, 1.396515869140625, 1.3962587890625, 1.3965823974609375, 1.3966663818359375, 1.396378662109375, 1.3987901611328124, 1.3961759033203125, 1.3963653564453125, 1.3963018798828124, 1.3964093017578125, 1.39652197265625, 1.3961544189453126, 1.3962332763671874, 1.39641650390625, 1.39623828125, 1.3962393798828125, 1.3961964111328125, 1.396336669921875, 1.3964564208984376, 1.3965137939453125, 1.3963756103515625, 1.3962701416015626, 1.396137939453125, 1.396569091796875, 1.39635205078125, 1.3963131103515625, 1.396252685546875, 1.3963427734375, 1.3963970947265625, 1.3995396728515626, 1.39681591796875, 1.396443115234375, 1.396485107421875, 1.3962608642578125, 1.3962506103515624, 1.3963192138671876, 1.3964769287109375, 1.396210693359375, 1.3968311767578125, 1.396547607421875, 1.39646875, 1.39649951171875, 1.396304931640625, 1.3964041748046876, 1.3961563720703125, 2.961944580078125, 1.39614208984375, 1.396427734375, 1.39610009765625, 1.3963775634765625, 1.397116943359375, 1.3965689697265624, 1.396379638671875, 1.396430908203125, 1.3963284912109375, 1.3966080322265626, 1.3961544189453126, 1.3964501953125, 1.396547607421875, 1.396279296875, 1.3963203125, 1.3963448486328125, 1.396454345703125, 1.3965128173828125, 1.396253662109375, 1.396316162109375, 1.396046875, 1.396062255859375, 1.3964974365234375, 1.396443115234375, 1.3962864990234376, 1.3960980224609374, 1.3964810791015625, 1.3962884521484376, 1.396199462890625, 1.396241455078125, 1.3960233154296875, 1.3961175537109376, 1.396675537109375, 1.3962762451171875, 1.3964237060546876, 1.3964083251953125, 1.3962884521484376, 1.3965762939453126, 1.3964708251953124, 1.396631591796875, 1.3964022216796874, 1.396336669921875, 1.396317138671875, 1.400958984375, 1.396559814453125, 1.3964820556640625, 1.39627001953125, 1.3966888427734374, 1.3963427734375, 1.396474853515625, 1.3962342529296874, 1.3963448486328125, 1.39620654296875, 1.396552734375, 1.3967205810546874, 1.3966029052734374, 1.3963560791015626, 1.3962332763671874, 1.3963212890625, 1.3964134521484375, 1.396279296875, 1.39620654296875, 2.96119921875, 1.3962393798828125, 1.39626806640625, 1.3964124755859375, 1.3966131591796875, 1.396216796875, 1.3965548095703124, 1.3961031494140625, 1.396580322265625, 1.396380615234375, 1.3969398193359375, 1.396273193359375, 1.39626904296875, 1.39641650390625, 1.3962547607421876, 1.39647900390625, 1.396178955078125, 1.3985556640625, 1.3966253662109376, 1.3968516845703125, 1.3966131591796875, 1.3960867919921875, 1.39594140625, 1.3963447265625, 1.396115478515625, 1.396316162109375, 1.3962301025390624, 1.3962618408203125, 1.3963406982421875, 1.3959669189453126, 1.396400146484375, 1.3961195068359376, 1.3963663330078124, 1.396368408203125, 1.3962496337890624, 1.39649951171875, 1.3963284912109375, 1.3964493408203125, 1.3962188720703126, 1.396115478515625, 1.3963602294921875, 1.3963663330078124, 1.3966182861328125, 1.3964237060546876, 1.3963038330078126, 1.3963653564453125, 1.39626806640625, 1.396369384765625, 1.3965137939453125, 1.3962486572265624, 1.3964317626953124, 1.3962711181640626, 1.39637353515625, 1.3964647216796875, 1.39634375, 1.3963253173828125, 1.396611083984375, 1.4011822509765626, 1.3966837158203125, 1.3966038818359374, 1.39641650390625, 1.3965537109375, 1.3965322265625, 2.960329833984375, 1.3966304931640625, 1.396526123046875, 1.3960919189453125, 1.396494384765625, 1.39626904296875, 1.3971036376953125, 1.39640625, 1.3964461669921875, 1.3961175537109376, 1.3965537109375, 1.3965025634765624, 1.3961400146484375, 1.396515869140625, 1.396490234375, 1.396189208984375, 1.3961563720703125, 1.3964564208984376, 1.3965404052734376, 1.396285400390625, 1.3961553955078125, 1.396074462890625, 1.395947509765625, 1.3964676513671874, 1.3965732421875, 1.3963448486328125, 1.3962977294921874, 1.3964267578125, 1.3964461669921875, 1.39627001953125, 1.3965732421875, 1.396262939453125, 1.39629052734375, 1.39640625, 1.3961185302734376, 1.3963243408203125, 1.3963826904296874, 1.3964442138671875, 1.400627197265625, 1.396809814453125, 1.396582275390625, 1.396357177734375, 1.39643798828125, 1.3963756103515625, 1.3962301025390624, 1.3963294677734375, 1.39649951171875, 1.3963336181640624, 1.3967288818359376, 1.396634521484375, 1.3963294677734375, 1.3961072998046875, 1.3963233642578126, 1.3963182373046874, 1.39650146484375, 1.396305908203125, 1.3965128173828125, 1.396368408203125, 1.3962659912109374, 1.39628955078125, 1.3963355712890626, 1.39620654296875, 1.39628955078125, 2.96319189453125, 1.39616259765625, 1.3965670166015625, 1.39634375, 1.39635205078125, 1.3960509033203126, 1.39610107421875, 1.3959833984375, 1.3964052734375, 1.3961861572265626, 1.39652197265625, 1.39647998046875, 1.3962496337890624, 1.3968414306640624, 1.3966519775390625, 1.3960438232421875, 1.3962802734375, 1.3961964111328125, 1.3965936279296876, 1.3964493408203125, 1.3962025146484376, 1.396211669921875, 1.3961871337890626, 1.3990369873046875, 1.39656396484375, 1.39658447265625, 1.3963192138671876, 1.3963756103515625, 1.396262939453125, 1.3963499755859374, 1.3965291748046875, 1.3964564208984376, 1.396116455078125, 1.3963765869140625, 1.396496337890625, 1.396341796875, 1.3961666259765626, 1.396453369140625, 1.396306884765625, 1.3964974365234375, 1.396357177734375, 1.396156494140625, 1.396138916015625, 1.39642578125, 1.3961502685546876, 1.396420654296875, 1.3963642578125, 1.39631201171875, 1.396749267578125, 1.396473876953125, 1.396526123046875, 1.396474853515625, 1.3964871826171874, 1.396432861328125, 1.396547607421875, 1.3964410400390626, 1.4002022705078125, 1.3966510009765625, 1.39624658203125, 1.396135986328125, 1.3961492919921874, 1.3963140869140624, 1.39603759765625, 2.96123388671875, 1.3961964111328125, 1.3963387451171876, 1.3962198486328126, 1.39618408203125, 1.3961973876953124, 1.3963714599609376, 1.396177978515625, 1.3965966796875, 1.396106201171875, 1.396715576171875, 1.3962823486328124, 1.3963294677734375, 1.3963028564453126, 1.396380615234375, 1.3961964111328125, 1.398497314453125, 1.3963970947265625, 1.3965679931640624, 1.396317138671875, 1.3965782470703125, 1.3963919677734375, 1.3962659912109374, 1.3962720947265626, 1.396201416015625, 1.396317138671875, 1.3962393798828125, 1.396137939453125, 1.3964708251953124, 1.3964288330078125, 1.3963857421875, 1.3960755615234375, 1.3963480224609375, 1.396232177734375, 1.3961451416015624, 1.3963182373046874, 1.396641845703125, 1.3966080322265626, 1.396378662109375, 1.396273193359375, 1.3963345947265624, 1.3962322998046874, 1.3966744384765626, 1.3964073486328126, 1.3963243408203125, 1.3962373046875, 1.3963499755859374, 1.3963345947265624, 1.3965465087890625, 1.396432861328125, 1.39626904296875, 1.39639404296875, 1.3961748046875, 1.3964892578125, 1.3966878662109374, 1.396232177734375, 1.396611083984375, 1.3963765869140625, 1.39654248046875, 1.39650048828125, 1.39634375, 1.3964093017578125, 1.3962454833984375, 2.96591357421875, 1.3961614990234374, 1.396864013671875, 1.3961820068359374, 1.396379638671875, 1.3964666748046874, 1.3964503173828124, 1.3963642578125, 1.39664892578125, 1.39633154296875, 1.396452392578125, 1.3962650146484374, 1.396241455078125, 1.3963756103515625, 1.3963775634765625, 1.396111328125, 1.3964503173828124, 1.3980078125, 1.3967994384765625, 1.396642822265625, 1.3961103515625, 1.3965987548828125, 1.396348876953125, 1.3962291259765625, 1.396262939453125, 1.39631201171875, 1.3963212890625, 1.3961759033203125, 1.3965260009765625, 1.3965496826171875, 1.396384765625, 1.3963182373046874, 1.396358154296875, 1.3962158203125, 1.3962720947265626, 1.396420654296875, 1.3962281494140625, 1.3967728271484374, 1.3964512939453124, 1.39666845703125, 1.3965936279296876, 1.3961728515625, 1.39624853515625, 1.3961861572265626, 1.39656396484375, 1.3964697265625, 1.396485107421875, 1.396537353515625, 1.3965670166015625, 1.3964073486328126, 1.3962578125, 1.3964390869140626, 1.3966990966796875, 1.39630078125, 1.3968045654296875, 1.3967421875, 1.3966285400390626, 1.3964482421875, 1.396533203125, 1.396506591796875, 1.396284423828125, 1.3962977294921874, 1.400701904296875, 2.960291748046875, 1.3964564208984376, 1.396536376953125, 1.396294677734375, 1.396358154296875, 1.39620458984375, 1.3963621826171875, 1.3960714111328125, 1.396537353515625, 1.396304931640625, 1.3964451904296875, 1.396430908203125, 1.3963642578125, 1.3963837890625, 1.396828125, 1.3963817138671875, 1.39660498046875, 1.396589599609375, 1.39631201171875, 1.396283447265625, 1.396177978515625, 1.39618505859375, 1.396168701171875, 1.396116455078125, 1.39642578125, 1.396654052734375, 1.399869384765625, 1.3961328125, 1.3965076904296876, 1.3962025146484376, 1.3964339599609374, 1.3961707763671876, 1.3962506103515624, 1.396093994140625, 1.3964124755859375, 1.396494384765625, 1.3964676513671874, 1.396410400390625, 1.396167724609375, 1.3962772216796875, 1.396236328125, 1.3962117919921875, 1.39633251953125, 1.396305908203125, 1.396552734375, 1.3963140869140624, 1.39652197265625, 1.396462646484375, 1.3968414306640624, 1.39652099609375, 1.396305908203125, 1.396357177734375, 1.3965721435546874, 1.39648193359375, 1.396537353515625, 1.396272216796875, 1.396579345703125, 1.3963602294921875, 1.3965322265625, 1.396306884765625, 1.3964041748046876, 1.396275146484375, 1.3961829833984376]",tokens/s,0.7047951516023033,, @@ -7047,7 +7047,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e9e9-204b01563abfaf961831988a;83d1bbd7-1c19-497b-8855-f4292831d35e) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667231d7-54b77c3367dba6884dbd011a;986c08e0-beb5-4521-93f6-3c1708c17ed1) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -7132,7 +7132,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e6cb-554fdb5b18f9644a0f1a2d88;bf5994c8-713f-491f-994e-bf458f659f0e) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722ef0-3e26f0fe76716c4b5814346d;e68da103-2102-40c8-8785-2176a6065960) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7194,7 +7194,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d7c4-7477b5f82140e208522b3b6c;7bb40a3d-ad51-4a14-b980-5a2a505c5c3e) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66722951-2c97737d2e57b2321b346ae6;428cd301-becc-48c5-9c01-87193181ed0c) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -7321,7 +7321,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e67b-30e88d9b17996be30fa0a4fb;783340d3-1135-40d5-a600-cbfe3899dc03) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722ea1-1131dc6973f534f36991535b;abee89c3-51fd-4b78-8f6d-06cc601d391a) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7394,7 +7394,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670d83d-2866593f423ab21f3e2a5af1;71ca5190-ada3-4d7b-8e72-1836ae030172) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667229cc-10400a7a6c88db2461ca2ce4;7765f68f-0642-4444-83f0-38bbf7c15419) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7453,7 +7453,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpq4pybn7x/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpr56gttln/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.1+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,B,B,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,False,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.217-205.860.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.2,,0.31.0,,,,1.20.0,,,,0.11.1,,"Traceback (most recent call last): @@ -7495,7 +7495,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e7e6-73f330ad76cdb97e33ae9f80;78b902c8-ec28-4237-bc48-028cd4d03d59) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6672300c-2ab4205f47cbddfb2f05991c;d32fc0f3-7aba-4d11-b486-61577c5ac693) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7552,7 +7552,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpx9wjvhnv/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: FalconForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpmm2p40jg/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,TencentARC/Mistral_Pro_8B_v0.1,TencentARC/Mistral_Pro_8B_v0.1,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,5937.815552,7053.246464,0.0,6406.79936,6098.60864,s,10,4.369107391357422,0.4369107391357422,0.0007208936334109947,0.43653796386718746,0.43786113586425784,0.4381150161743164,0.4383181204223633,"[0.438368896484375, 0.43641900634765624, 0.43653814697265625, 0.43647592163085935, 0.43767733764648437, 0.43614486694335936, 0.43780471801757814, 0.43653778076171873, 0.4368502807617187, 0.4362904357910156]",tokens/s,585.9320384442743,kWh,5.155067728913349e-06,2.8244480818023514e-06,2.4793715487132108e-05,3.2773231297847804e-05,tokens/kWh,7811252.960485814,MB,5937.815552,7055.343616,0.0,6408.896512,6157.952,s,10,245.01072265624998,24.501072265624998,0.0034318484925660374,24.5004931640625,24.5048515625,24.5070751953125,24.508854101562502,"[24.50105078125, 24.49864453125, 24.50170703125, 24.49968359375, 24.496791015625, 24.509298828125, 24.50153515625, 24.504357421875, 24.499935546875, 24.49771875]",tokens/s,2.571316035355277,kWh,0.0002892534986599056,0.000158535909402625,0.0013504422735896686,0.0017982316816521993,tokens/kWh,35034.417779869254,,s,629,248.55180075073253,0.3951538962650754,0.05191103267325261,0.38883840942382814,0.3892619018554688,0.38950277099609376,0.8256093530273438,"[0.38878106689453124, 0.38896536254882813, 0.388632568359375, 0.38873394775390624, 0.38894488525390625, 0.3886663818359375, 0.3890759582519531, 0.3891517333984375, 0.3889776611328125, 0.38876263427734375, 0.38862646484375, 0.38881890869140623, 0.3890196533203125, 0.3885875244140625, 0.38878616333007815, 0.3895060424804688, 0.38884146118164065, 0.38903399658203125, 0.389644287109375, 0.38915277099609374, 0.3891998596191406, 0.3889284973144531, 0.38931558227539065, 0.3889561462402344, 0.3889131469726563, 0.38835916137695314, 0.38861004638671875, 0.38877081298828126, 0.388674560546875, 0.3886622619628906, 0.3889428405761719, 0.38882098388671876, 0.3894353942871094, 0.3886909484863281, 0.38898175048828126, 0.38895718383789063, 0.38856500244140624, 0.3886120910644531, 0.389538818359375, 0.38871450805664065, 0.38864486694335937, 0.3886510009765625, 0.3886561279296875, 0.38851071166992185, 0.3891835021972656, 0.38876364135742186, 0.3887032470703125, 0.3888701477050781, 0.38866943359375, 0.3892899780273438, 0.3888373718261719, 0.388600830078125, 0.3887615966796875, 0.3901071472167969, 0.3886766052246094, 0.3887728576660156, 0.3887442016601563, 0.3886929931640625, 0.3890124816894531, 0.3890810852050781, 0.38887628173828126, 0.388568115234375, 0.8254780883789062, 0.3889141845703125, 0.38875750732421877, 0.38877593994140625, 0.3889254455566406, 0.3885977478027344, 0.3893872680664062, 0.38873907470703123, 0.38861721801757815, 0.3886346130371094, 0.38904217529296875, 0.3888578491210937, 0.38904013061523435, 0.38923876953125, 0.38863873291015627, 0.38872164916992186, 0.38861822509765626, 0.3884800109863281, 0.3887462463378906, 0.3886069641113281, 0.3893196716308594, 0.38932070922851564, 0.3890186157226562, 0.3887646789550781, 0.38994842529296875, 0.38882098388671876, 0.3890513916015625, 0.3891650695800781, 0.38888037109375, 0.3886991882324219, 0.3889468688964844, 0.3885434875488281, 0.3888199768066406, 0.3885844421386719, 0.388600830078125, 0.3886417846679687, 0.3889479675292969, 0.3886417846679687, 0.3886336059570312, 0.3887790222167969, 0.38892236328125, 0.38850970458984374, 0.38881793212890625, 0.3887523803710938, 0.3886868591308594, 0.38856500244140624, 0.3889899597167969, 0.38897457885742187, 0.38872164916992186, 0.3891282043457031, 0.38919577026367186, 0.3890595703125, 0.3892817993164063, 0.38935140991210937, 0.3886397399902344, 0.38855474853515626, 0.3887595520019531, 0.3887523803710938, 0.3889899597167969, 0.38894900512695313, 0.38868582153320314, 0.38889370727539063, 0.3889704895019531, 0.8256952514648438, 0.38874215698242187, 0.38912716674804687, 0.3890257873535156, 0.38883328247070315, 0.3886970825195312, 0.3889541015625, 0.38887527465820315, 0.389212158203125, 0.389137451171875, 0.38895819091796874, 0.3886970825195312, 0.3888670654296875, 0.38879641723632813, 0.38865716552734375, 0.3888598937988281, 0.3887564697265625, 0.3886991271972656, 0.38894900512695313, 0.3889837951660156, 0.38897357177734376, 0.3888793640136719, 0.38878106689453124, 0.38885479736328127, 0.3889776611328125, 0.38883840942382814, 0.38856396484375, 0.3887400817871094, 0.38860800170898435, 0.3888199768066406, 0.38868582153320314, 0.3887298583984375, 0.38889166259765623, 0.38933505249023437, 0.3892725830078125, 0.3889756164550781, 0.38903604125976565, 0.38886810302734376, 0.38863665771484374, 0.3886458740234375, 0.3889704895019531, 0.38879129028320314, 0.3888148498535156, 0.3887943725585937, 0.3887175598144531, 0.38925106811523436, 0.3889131469726563, 0.38866534423828125, 0.3887677307128906, 0.38883840942382814, 0.388853759765625, 0.388853759765625, 0.3888025512695312, 0.38896536254882813, 0.38888958740234375, 0.3887513732910156, 0.38878823852539063, 0.38919064331054687, 0.38868582153320314, 0.39011123657226565, 0.3893196716308594, 0.38929306030273436, 0.38961663818359377, 0.8260413208007813, 0.38885992431640626, 0.3889643249511719, 0.3887523803710938, 0.3887247314453125, 0.3887185974121094, 0.38895718383789063, 0.38912408447265623, 0.3887984619140625, 0.3889039306640625, 0.388822021484375, 0.3887298583984375, 0.388811767578125, 0.3887718505859375, 0.38874215698242187, 0.3888404541015625, 0.38900531005859373, 0.38877694702148435, 0.3888035888671875, 0.3886673889160156, 0.38864999389648436, 0.38919064331054687, 0.38883941650390624, 0.38895718383789063, 0.38876263427734375, 0.38873394775390624, 0.3887442016601563, 0.3888988037109375, 0.38905035400390625, 0.38884658813476564, 0.38873294067382813, 0.3889889221191406, 0.3889121398925781, 0.3887595520019531, 0.3887544250488281, 0.38894488525390625, 0.38860800170898435, 0.38863873291015627, 0.3886981201171875, 0.38873294067382813, 0.3888630065917969, 0.3888568115234375, 0.3888783264160156, 0.38862539672851565, 0.3889613037109375, 0.389233642578125, 0.38876263427734375, 0.38913946533203125, 0.3891650695800781, 0.3887237243652344, 0.38868582153320314, 0.38899713134765623, 0.3889704895019531, 0.3898193969726563, 0.3890257873535156, 0.3886673889160156, 0.38894692993164065, 0.3890390930175781, 0.3892991943359375, 0.38878515625, 0.3889059753417969, 0.3887462463378906, 0.3888424987792969, 0.826693603515625, 0.3888025512695312, 0.3887523803710938, 0.3885926513671875, 0.3887923278808594, 0.3886120910644531, 0.38872576904296874, 0.38866021728515626, 0.3889807434082031, 0.38920294189453125, 0.38865304565429687, 0.38867559814453123, 0.3889039306640625, 0.388706298828125, 0.3886264343261719, 0.3890257873535156, 0.3886796875, 0.38881893920898436, 0.388748291015625, 0.3887544250488281, 0.3898101806640625, 0.38882098388671876, 0.3889725341796875, 0.38886605834960936, 0.3887820739746094, 0.38895001220703124, 0.3888015441894531, 0.38883840942382814, 0.3887093811035156, 0.38877593994140625, 0.3885823974609375, 0.3885844421386719, 0.3888087158203125, 0.388632568359375, 0.388895751953125, 0.3889407958984375, 0.3887400817871094, 0.38869403076171877, 0.38845440673828124, 0.388706298828125, 0.3888496704101562, 0.38918142700195313, 0.3888035888671875, 0.38866534423828125, 0.3889039306640625, 0.38894488525390625, 0.3888486328125, 0.38889984130859373, 0.3886970825195312, 0.3889612731933594, 0.3889326171875, 0.38907904052734377, 0.38912612915039063, 0.388790283203125, 0.3889459228515625, 0.38889675903320314, 0.3888025512695312, 0.38870016479492187, 0.38883428955078125, 0.3888517150878906, 0.38923776245117186, 0.38900222778320315, 0.38878616333007815, 0.8254668579101563, 0.389001220703125, 0.3892408447265625, 0.38915377807617185, 0.38880767822265627, 0.38871142578125, 0.38886297607421877, 0.3891148681640625, 0.38876364135742186, 0.38882720947265625, 0.3886867980957031, 0.3885404052734375, 0.38891827392578127, 0.38905035400390625, 0.38921829223632814, 0.3889141845703125, 0.3891558532714844, 0.38889471435546874, 0.3890810852050781, 0.38909234619140626, 0.3887319030761719, 0.38957876586914064, 0.3891712036132812, 0.3887923583984375, 0.3895316162109375, 0.38924493408203126, 0.38886605834960936, 0.38888653564453124, 0.38975283813476563, 0.38893875122070315, 0.38879641723632813, 0.38891009521484377, 0.38910565185546875, 0.3892593078613281, 0.39087921142578125, 0.38871551513671876, 0.38870834350585937, 0.38860186767578125, 0.3886960754394531, 0.38911077880859374, 0.38857318115234374, 0.3887657470703125, 0.3889714660644531, 0.38878106689453124, 0.3889807434082031, 0.3888148498535156, 0.3887872009277344, 0.3890739135742188, 0.38941082763671875, 0.3891589050292969, 0.38924798583984377, 0.3888977966308594, 0.38902792358398436, 0.3890799560546875, 0.3894364013671875, 0.3890677795410156, 0.388959228515625, 0.38916094970703125, 0.3891568603515625, 0.3888988037109375, 0.38902987670898437, 0.38889370727539063, 0.38949786376953127, 0.8270551147460937, 0.38860186767578125, 0.38860595703125, 0.38868377685546873, 0.38870834350585937, 0.38866021728515626, 0.3888844909667969, 0.38888858032226564, 0.3890831298828125, 0.388947998046875, 0.38878512573242185, 0.389180419921875, 0.38898175048828126, 0.38901556396484377, 0.38881381225585937, 0.38938214111328123, 0.38876568603515627, 0.38882098388671876, 0.3885455322265625, 0.3887093811035156, 0.3890534362792969, 0.38878515625, 0.3887400817871094, 0.3889725341796875, 0.3888455810546875, 0.38887628173828126, 0.3890677795410156, 0.38898175048828126, 0.38893771362304685, 0.3888015441894531, 0.38900735473632814, 0.3888015441894531, 0.38892440795898436, 0.3890124816894531, 0.3895828552246094, 0.3893197021484375, 0.3891650390625, 0.3890083923339844, 0.38891827392578127, 0.38878311157226564, 0.3886766052246094, 0.38947021484375, 0.3891251220703125, 0.38870834350585937, 0.3887933349609375, 0.3888517150878906, 0.38911181640625, 0.38871142578125, 0.38868582153320314, 0.3887319030761719, 0.388611083984375, 0.38874215698242187, 0.388890625, 0.3886889038085938, 0.38894692993164065, 0.38870220947265627, 0.3885906066894531, 0.38906878662109373, 0.3886192626953125, 0.38860186767578125, 0.3894384765625, 0.38954290771484373, 0.3893800964355469, 0.825660400390625, 0.38914047241210936, 0.3887677307128906, 0.38940365600585936, 0.38903704833984376, 0.3888568420410156, 0.38899917602539064, 0.3896258544921875, 0.3887247314453125, 0.3890565185546875, 0.3885342712402344, 0.38850048828125, 0.38885888671875, 0.38866021728515626, 0.38881793212890625, 0.3884103698730469, 0.38903604125976565, 0.38858648681640623, 0.3889070129394531, 0.3887185974121094, 0.38888858032226564, 0.3892193298339844, 0.3894497375488281, 0.38906060791015623, 0.3886120910644531, 0.388600830078125, 0.38877490234375, 0.3888005065917969, 0.38874725341796873, 0.38922955322265623, 0.3888926696777344, 0.3885957336425781, 0.38874212646484374, 0.3889346618652344, 0.3888650207519531, 0.38876876831054685, 0.388664306640625, 0.38930841064453126, 0.3888230285644531, 0.38908517456054686, 0.38917633056640627, 0.38882611083984375, 0.3889510498046875, 0.38876058959960935, 0.3885772705078125, 0.3890186157226562, 0.3886346130371094, 0.3894353942871094, 0.389170166015625, 0.3888036193847656, 0.38960125732421874, 0.3888005065917969, 0.3895419006347656, 0.3890739135742188, 0.3897927551269531, 0.38913229370117186, 0.38885888671875, 0.3889837951660156, 0.388969482421875, 0.38935653686523436, 0.38904013061523435, 0.3888988037109375, 0.38913433837890626, 0.8261171264648437, 0.38934527587890627, 0.3890943908691406, 0.3892561950683594, 0.38904730224609374, 0.3889776611328125, 0.38878823852539063, 0.38880459594726563, 0.38874215698242187, 0.38939239501953127, 0.3890657348632813, 0.3885916137695313, 0.38864999389648436, 0.3887585144042969, 0.388790283203125, 0.3887032470703125, 0.3896012878417969, 0.3887442016601563, 0.3886336059570312, 0.38861004638671875, 0.389037109375, 0.3888260498046875, 0.388701171875, 0.38865304565429687, 0.38868069458007815, 0.38857318115234374, 0.3886213073730469, 0.3887431640625, 0.38874725341796873, 0.38863873291015627, 0.388595703125, 0.38875341796875, 0.38857522583007814, 0.3887923278808594, 0.38920806884765624, 0.3891712036132812, 0.38883123779296874, 0.3889407958984375, 0.388748291015625, 0.3887503356933594, 0.38887527465820315, 0.38860287475585936, 0.38913742065429685, 0.3887790222167969, 0.3885629577636719, 0.38873394775390624, 0.38880563354492187, 0.38863565063476563, 0.38857318115234374, 0.3885987854003906, 0.38900531005859373, 0.3888097229003906, 0.38852197265625, 0.3885834350585938, 0.38872576904296874, 0.38862234497070314, 0.3915376586914063, 0.38877694702148435, 0.3889428405761719, 0.38899917602539064, 0.38912716674804687, 0.38925726318359377, 0.3891875610351562, 0.8258375854492187, 0.388811767578125, 0.38908212280273435, 0.3887953796386719, 0.38926849365234373, 0.38856192016601565, 0.38865304565429687, 0.38886297607421877, 0.38894692993164065, 0.38873086547851565, 0.3887933349609375, 0.3884656677246094, 0.38854656982421876, 0.38860287475585936, 0.3886929931640625, 0.3886868591308594, 0.38878411865234375, 0.3888066711425781, 0.3886141357421875, 0.3886141357421875, 0.38892340087890626, 0.38862335205078125, 0.3893319702148437, 0.3891773376464844, 0.38893875122070315, 0.38863565063476563, 0.3889203186035156, 0.3887298583984375, 0.38881689453125, 0.3891302490234375, 0.3888824462890625, 0.3886489562988281, 0.3886243896484375, 0.3889407958984375, 0.38874111938476563, 0.38874111938476563, 0.38891009521484377, 0.3886141357421875, 0.38878823852539063, 0.38871142578125, 0.3888291931152344, 0.38855270385742186, 0.38883328247070315, 0.388885498046875, 0.38914764404296875, 0.38934017944335936, 0.38926025390625, 0.38955938720703126, 0.3897701416015625, 0.3890442199707031, 0.38865716552734375, 0.3886612548828125, 0.3888455810546875, 0.38867559814453123, 0.38898483276367185, 0.3890862121582031, 0.38879129028320314, 0.38848614501953127, 0.3886970825195312, 0.3888097229003906, 0.3888128356933594, 0.3887206115722656, 0.38876980590820315]",tokens/s,2.530659597315939,, @@ -7595,7 +7595,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e494-46292da06990fbe7490ac973;ee31bf9d-5f1f-4982-9e82-7ee98c875a61) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722ce0-31d90480295083712b9a427f;a9948a6a-80a8-4b42-8135-88764ce1890b) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7653,7 +7653,7 @@ ChildProcessError: Traceback (most recent call last): cls._check_and_enable_flash_attn_2( File ""/usr/local/lib/python3.10/dist-packages/transformers/modeling_utils.py"", line 1560, in _check_and_enable_flash_attn_2 raise ValueError( -ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmphgmks7nm/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new +ValueError: XGLMForCausalLM does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co//tmp/tmpfz_t_pbn/no_weights_model/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new ",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, 4bit-gptq-exllama-v2-flash_attention_2,pytorch,2.3.0+cu121,optimum_benchmark.backends.pytorch.backend.PyTorchBackend,text-generation,transformers,facebook/opt-350m,facebook/opt-350m,cuda,0,42,,,True,True,True,True,,float16,True,False,,flash_attention_2,,False,,False,forward,gptq,4,True,2,256,False,,inference,optimum_benchmark.scenarios.inference.scenario.InferenceScenario,10,10,10,1,2,256,,True,True,True,64,64,process,optimum_benchmark.launchers.process.launcher.ProcessLauncher,True,kill,,spawn, AMD EPYC 7R32,16,66697.29792,Linux,x86_64,Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35,x86_64,3.10.12,['NVIDIA A10G'],1,24146608128,0.2.1,,4.41.0,,0.30.1,,,,1.19.2,,,,0.11.1,,,,,,MB,1269.682176,1019.74016,0.0,373.293056,277.410816,s,10,0.37923129653930665,0.037923129653930666,0.0012152886645715089,0.038048974990844725,0.038519168853759764,0.0396308479309082,0.04052019119262695,"[0.04074252700805664, 0.03821686553955078, 0.03797974395751953, 0.03823471832275391, 0.03626841735839844, 0.03608185577392578, 0.03733708953857422, 0.03811382293701172, 0.03827212905883789, 0.03798412704467773]",tokens/s,6750.497713035296,kWh,4.301884120160883e-07,2.3572264180256752e-07,1.164744366137984e-06,1.83065541995664e-06,tokens/kWh,139840626.0453229,MB,1269.682176,1019.74016,0.0,373.293056,323.047424,s,10,23.493592773437502,2.3493592773437504,0.014632907695307939,2.3561588134765623,2.362567529296875,2.3635712402343754,2.3643742089843753,"[2.34352734375, 2.36090478515625, 2.362344482421875, 2.364574951171875, 2.338309326171875, 2.331607421875, 2.357506103515625, 2.3608046875, 2.3548115234375, 2.3192021484375]",tokens/s,26.81582191687153,kWh,2.642717914328431e-05,1.4482863030778391e-05,6.568572578086276e-05,0.00010659576795492543,tokens/kWh,591017.8350292469,,s,629,23.794291763305676,0.037828762739754634,0.004511848848474209,0.03739033508300781,0.03767849044799805,0.03791810684204102,0.07427948760986337,"[0.03534745788574219, 0.03654348754882813, 0.03639910507202149, 0.03678003311157227, 0.03678003311157227, 0.036736000061035154, 0.036125694274902344, 0.03654348754882813, 0.036519935607910156, 0.0366110725402832, 0.03650764846801758, 0.03670425415039062, 0.03646771240234375, 0.037282817840576174, 0.03782860946655273, 0.03862527847290039, 0.037749759674072264, 0.03757056045532227, 0.037465087890625, 0.0372408332824707, 0.03734630584716797, 0.03750092697143555, 0.03727872085571289, 0.03741388702392578, 0.03740262222290039, 0.03726131057739258, 0.037303295135498044, 0.03731353759765625, 0.037188640594482424, 0.037268447875976565, 0.03742617416381836, 0.03730124664306641, 0.03739136123657227, 0.037250049591064455, 0.03735039901733399, 0.03733913421630859, 0.03732787322998047, 0.03716505432128906, 0.03726233673095703, 0.0372490234375, 0.03732070541381836, 0.03729305648803711, 0.03743129730224609, 0.03723980712890625, 0.037269504547119144, 0.03735039901733399, 0.03727769470214844, 0.03724800109863281, 0.03734124755859375, 0.037339073181152344, 0.037269504547119144, 0.03736678314208984, 0.0371701774597168, 0.037266433715820314, 0.03737190246582031, 0.037318656921386716, 0.03726847839355469, 0.03745894241333008, 0.03732787322998047, 0.03729715347290039, 0.037501953125, 0.037501953125, 0.07589580535888672, 0.03728793716430664, 0.037407745361328126, 0.037207038879394534, 0.03731353759765625, 0.03784089660644531, 0.037353473663330077, 0.03739340972900391, 0.0374015998840332, 0.03751731109619141, 0.03856486511230469, 0.037754878997802735, 0.037369857788085936, 0.03734220886230469, 0.03731353759765625, 0.0373125114440918, 0.03733606338500976, 0.037425151824951174, 0.037303295135498044, 0.037362686157226564, 0.038662143707275394, 0.037748737335205076, 0.03751116943359375, 0.03752243041992188, 0.03749785614013672, 0.03739340972900391, 0.03728793716430664, 0.03726540756225586, 0.03750707244873047, 0.03739340972900391, 0.03738828659057617, 0.03735859298706055, 0.037395454406738284, 0.03731558227539063, 0.037868545532226565, 0.037456897735595705, 0.03736883163452148, 0.03723161697387695, 0.03758489608764649, 0.03761151885986328, 0.03742822265625, 0.037384193420410154, 0.03741900634765625, 0.03744976043701172, 0.03728688049316406, 0.03746815872192383, 0.03732275390625, 0.037525505065917966, 0.0375203857421875, 0.03739340972900391, 0.03733913421630859, 0.03781631851196289, 0.03749785614013672, 0.03746815872192383, 0.03744870376586914, 0.037324798583984374, 0.03752345657348633, 0.03728691101074219, 0.0374015998840332, 0.03738828659057617, 0.037454849243164064, 0.037454849243164064, 0.03749273681640625, 0.0757176284790039, 0.03736166381835938, 0.03737702560424805, 0.03760332870483398, 0.03744768142700195, 0.037354496002197264, 0.03724595260620117, 0.037343231201171875, 0.03800678253173828, 0.03777740859985351, 0.03753472137451172, 0.03747430419921875, 0.0374835205078125, 0.037585918426513674, 0.037338111877441404, 0.0374466552734375, 0.03750912094116211, 0.037379070281982424, 0.03765964889526367, 0.03735551834106445, 0.03742310333251953, 0.03738726425170898, 0.03740364837646484, 0.037495807647705076, 0.03738214492797851, 0.037310462951660156, 0.03733401489257813, 0.03747430419921875, 0.03743436813354492, 0.0373831672668457, 0.03754393768310547, 0.037370880126953124, 0.03760025787353516, 0.037601280212402347, 0.037510143280029294, 0.0373043212890625, 0.0373125114440918, 0.03747225570678711, 0.03762688064575195, 0.037282817840576174, 0.037294078826904296, 0.03753267288208008, 0.037607425689697264, 0.037400577545166014, 0.037495807647705076, 0.03745075225830078, 0.038043647766113284, 0.037773311614990236, 0.03741798400878906, 0.03741183853149414, 0.03771289443969727, 0.03767193603515625, 0.03745280075073242, 0.0372674560546875, 0.03750092697143555, 0.038141952514648435, 0.037969921112060545, 0.037384193420410154, 0.03739852905273437, 0.03727977752685547, 0.03753366470336914, 0.037572608947753904, 0.03738726425170898, 0.07624192047119141, 0.03899084854125977, 0.037817344665527344, 0.03795356750488281, 0.03747836685180664, 0.03777536010742188, 0.03754905700683594, 0.03737497711181641, 0.03731763076782227, 0.03754598236083984, 0.0381317138671875, 0.037395454406738284, 0.03748556900024414, 0.03738726425170898, 0.037493759155273435, 0.037318656921386716, 0.03763916778564453, 0.03733299255371094, 0.037407745361328126, 0.03729817581176758, 0.03746918487548828, 0.03743129730224609, 0.03740364837646484, 0.03732275390625, 0.03753267288208008, 0.03765350341796875, 0.03748863983154297, 0.03742617416381836, 0.03767708969116211, 0.03730940628051758, 0.03756748962402344, 0.03845017623901367, 0.03810201644897461, 0.03751321411132812, 0.03726847839355469, 0.03728998565673828, 0.03749273681640625, 0.03731148910522461, 0.03733606338500976, 0.03725107192993164, 0.03753881454467774, 0.03738726425170898, 0.037282817840576174, 0.03728179168701172, 0.0374917106628418, 0.037351425170898435, 0.03746819305419922, 0.037434337615966796, 0.037440513610839846, 0.03745587158203125, 0.03747532653808594, 0.03769651031494141, 0.03751935958862305, 0.037397502899169925, 0.03741491317749023, 0.037466110229492186, 0.03749478530883789, 0.037566463470458986, 0.03733401489257813, 0.03742822265625, 0.03756851196289063, 0.037501953125, 0.03759820938110352, 0.07184793853759766, 0.03510374450683594, 0.03521535873413086, 0.03516928100585937, 0.03511603164672852, 0.03513241577148438, 0.03520000076293945, 0.035148799896240236, 0.03514470291137695, 0.03511500930786133, 0.0367718391418457, 0.03750400161743164, 0.03728998565673828, 0.037482494354248046, 0.0374835205078125, 0.03756032180786133, 0.037303295135498044, 0.037424129486083986, 0.03768729782104492, 0.03739648056030274, 0.03758492660522461, 0.03751830291748047, 0.03748659133911133, 0.03748863983154297, 0.037410816192626956, 0.03736883163452148, 0.03754396820068359, 0.03754390335083008, 0.037591041564941405, 0.03752140808105469, 0.03739136123657227, 0.03755212783813477, 0.03739340972900391, 0.03741491317749023, 0.03769139099121094, 0.037408767700195314, 0.038091777801513675, 0.03745382308959961, 0.0376360969543457, 0.03739136123657227, 0.03732070541381836, 0.03746303939819336, 0.03743334579467773, 0.037364734649658206, 0.037526527404785154, 0.0374015998840332, 0.037416961669921874, 0.037416961669921874, 0.03750604629516602, 0.037427200317382815, 0.037367809295654295, 0.037498878479003905, 0.037397502899169925, 0.03749683380126953, 0.037622814178466794, 0.03773948669433594, 0.037416961669921874, 0.03744464111328125, 0.03751728057861328, 0.0374835205078125, 0.03757056045532227, 0.037477375030517575, 0.037720062255859374, 0.0716236801147461, 0.035062782287597655, 0.03518668746948242, 0.035019775390625, 0.03531980895996094, 0.035253246307373046, 0.03521843338012695, 0.034840576171875, 0.03494911956787109, 0.03526863861083984, 0.03520406341552734, 0.035253246307373046, 0.03508428955078125, 0.0372490234375, 0.037318656921386716, 0.03726335906982422, 0.03732787322998047, 0.03786342239379883, 0.037288959503173826, 0.0374835205078125, 0.03732992172241211, 0.03781631851196289, 0.038765567779541016, 0.03768012619018555, 0.03770982360839844, 0.037367809295654295, 0.03745894241333008, 0.03721318435668945, 0.03721420669555664, 0.03736678314208984, 0.03769343948364258, 0.03735756683349609, 0.03937484741210937, 0.03793203353881836, 0.037556224822998044, 0.03736883163452148, 0.03729510498046875, 0.03745177459716797, 0.03745792007446289, 0.037561344146728515, 0.037422080993652344, 0.037353473663330077, 0.037594112396240234, 0.03732070541381836, 0.03734732818603516, 0.037340160369873046, 0.037397502899169925, 0.037362686157226564, 0.03757875061035156, 0.03736064147949219, 0.037321727752685545, 0.03742822265625, 0.03740364837646484, 0.037318656921386716, 0.03724492645263672, 0.03730022430419922, 0.03753472137451172, 0.0374200325012207, 0.03728486251831055, 0.03722751998901367, 0.03739136123657227, 0.037362686157226564, 0.037375999450683595, 0.07522509002685547, 0.03725209426879883, 0.03725823974609375, 0.0374466552734375, 0.03733916854858398, 0.03731145477294922, 0.03730124664306641, 0.03746406555175781, 0.037375999450683595, 0.03750707244873047, 0.037292030334472655, 0.037459968566894535, 0.03742924880981445, 0.037405696868896485, 0.03738009643554688, 0.0372592658996582, 0.03754905700683594, 0.03736067199707031, 0.03739644622802735, 0.037394432067871096, 0.037394432067871096, 0.037392383575439454, 0.037389312744140625, 0.0372674560546875, 0.03746713638305664, 0.037416961669921874, 0.03733606338500976, 0.03751321411132812, 0.03757056045532227, 0.037351425170898435, 0.03736883163452148, 0.037335041046142575, 0.03735756683349609, 0.03738521575927734, 0.03734630584716797, 0.03721932983398438, 0.037397502899169925, 0.03740364837646484, 0.037465087890625, 0.03727974319458008, 0.03751116943359375, 0.03736883163452148, 0.03720294570922852, 0.03728179168701172, 0.03764019012451172, 0.037395454406738284, 0.037294078826904296, 0.037379070281982424, 0.03743231964111328, 0.03737705612182617, 0.03722953414916992, 0.03727360153198242, 0.03775692749023438, 0.03735039901733399, 0.03786652755737305, 0.03738211059570312, 0.03764019012451172, 0.03744255828857422, 0.03744153594970703, 0.03740671920776367, 0.038042625427246096, 0.03756032180786133, 0.03753267288208008, 0.07615692901611328, 0.03728486251831055, 0.03751628875732422, 0.03750092697143555, 0.03723161697387695, 0.03742105484008789, 0.03728384017944336, 0.03728793716430664, 0.03723369598388672, 0.037351390838623044, 0.0372592658996582, 0.03703603363037109, 0.03736576080322265, 0.03826483154296875, 0.03825356674194336, 0.037498878479003905, 0.03745075225830078, 0.03750604629516602, 0.03733606338500976, 0.0373493766784668, 0.037405696868896485, 0.037548030853271484, 0.037897216796875, 0.0374015998840332, 0.03749273681640625, 0.03787980651855469, 0.03769548797607422, 0.03733401489257813, 0.037424129486083986, 0.03724288177490234, 0.03732275390625, 0.03731353759765625, 0.037348350524902346, 0.0373043212890625, 0.03746099090576172, 0.03751731109619141, 0.037424129486083986, 0.037477375030517575, 0.03738623809814453, 0.037395454406738284, 0.0372828483581543, 0.0373001937866211, 0.03731455993652344, 0.03749683380126953, 0.037678081512451174, 0.03893964767456055, 0.037498878479003905, 0.03726847839355469, 0.03724595260620117, 0.037381118774414065, 0.03728998565673828, 0.03745177459716797, 0.03747635269165039, 0.037425151824951174, 0.03755724716186523, 0.03739033508300781, 0.03731763076782227, 0.03747532653808594, 0.03732070541381836, 0.03763302230834961, 0.0374917106628418, 0.03746303939819336, 0.037392383575439454, 0.07604838562011719, 0.03738726425170898, 0.03732275390625, 0.037763072967529294, 0.03760537719726562, 0.037493759155273435, 0.03744153594970703, 0.037367809295654295, 0.037353473663330077, 0.03748966217041016, 0.037321727752685545, 0.03739136123657227, 0.03747020721435547, 0.03746201705932617, 0.03758899307250976, 0.037288959503173826, 0.03742617416381836, 0.037359615325927735, 0.03741900634765625, 0.03733708953857422, 0.0372592658996582, 0.03737497711181641, 0.03758182525634766, 0.03737395095825195, 0.03727052688598633, 0.03736678314208984, 0.037477375030517575, 0.03746201705932617, 0.03737702560424805, 0.03690291213989258, 0.03730022430419922, 0.037541889190673826, 0.03752140808105469, 0.03749683380126953, 0.03740979385375977, 0.037354496002197264, 0.037292030334472655, 0.03724288177490234, 0.03722956848144531, 0.03722956848144531, 0.0375203857421875, 0.03734630584716797, 0.03717836761474609, 0.0371671028137207, 0.037372928619384765, 0.037341182708740234, 0.03730636978149414, 0.03713945770263672, 0.03719987106323242, 0.03745177459716797, 0.03732889556884766, 0.03736166381835938, 0.0373493766784668, 0.03739648056030274, 0.03730739212036133, 0.03730739212036133, 0.03730944061279297, 0.03724390411376953, 0.037353473663330077, 0.03740467071533203, 0.0375203857421875, 0.03734630584716797, 0.03746406555175781, 0.07565721893310547, 0.0372674560546875, 0.0373125114440918, 0.037479423522949216, 0.03725721740722656, 0.03723468780517578, 0.0374200325012207, 0.03733196640014649, 0.03729817581176758, 0.03751321411132812, 0.03730739212036133, 0.03743334579467773, 0.03721932983398438, 0.03751833724975586, 0.03735244750976562, 0.037408767700195314, 0.03565260696411133, 0.035312641143798826, 0.03527782440185547, 0.0352542724609375, 0.03772415924072266, 0.0377077751159668, 0.03799244689941406, 0.03948953628540039, 0.037106689453125, 0.03671449661254883, 0.03655372619628906, 0.03666534423828125, 0.03620249557495117, 0.03514470291137695, 0.03519692611694336, 0.035335166931152344, 0.03523788833618164, 0.03600076675415039, 0.03778662490844727, 0.03708927917480469, 0.03679846572875976, 0.036724769592285156, 0.03671651077270508, 0.0366192626953125, 0.036624416351318356, 0.03651375961303711, 0.03685990524291992, 0.03674524688720703, 0.03836924743652344, 0.03679436874389649, 0.03653734588623047, 0.03676671981811523, 0.03650559997558594, 0.03653529739379883, 0.036706302642822264, 0.03682099151611328, 0.036528129577636716, 0.03685887908935547, 0.03667865753173828, 0.036674560546875, 0.036618240356445314, 0.03663052749633789, 0.03661721420288086, 0.03660595321655274, 0.0365926399230957, 0.036544513702392575, 0.036792320251464845]",tokens/s,26.434911627418618,, @@ -7817,7 +7817,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e438-246e3a412863bd7a132be354;6a3c9916-c47c-4ea3-ba3e-24ebb097403f) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722c8f-3376dc4e13cd642a325b857e;d00be0be-8212-4140-97bb-e12cd4818dcb) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7887,7 +7887,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e8de-4557534862895c7557a696dc;aa30eafc-1c6d-4e6a-8963-9f656a28d49d) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667230d4-04a9b7e66f4594fe0c82a977;7627918e-cf08-4b13-b709-2eb4773ddbed) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -7957,7 +7957,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e879-78710c6c66c07dea5cdebb7b;98ac9079-ee4d-42f5-9fef-d049796cdf26) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66723084-291bfc6e12f0307a6c8b8959;fd7b30a8-3364-4bcf-b74d-db6b72cc7a93) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -8057,7 +8057,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e723-4616e4c66e7d031e35295179;aa9e5175-d162-44eb-aef6-b44e9f056b89) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722f47-0d8e60e60c8cc3c564f9831f;2b504ecc-790d-4093-906e-bb38e1d2f1b2) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -8128,7 +8128,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e79c-28a2aaf333ecf70349bb5bcf;50fc0d7e-7011-4a6b-96d1-6a0f82032036) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722fc1-22d0f1f70083e42c0d4a5fb8;1bfe263e-5284-4057-b4d0-c9ed4a964cd3) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -8312,7 +8312,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d7f3-418d75973fe823d314735444;9833771b-a751-46fe-bb87-8c895259c2ad) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66722981-6c3b5b6119cf85b24ade41f5;518bd66e-32a3-40ac-8141-c8ffdccb1678) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -8401,7 +8401,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e74b-577737943636402d0aa5eff9;cab60354-6b3b-4502-b086-923eff856bb7) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722f6e-6fe3bfe313a524b957f4982b;09fc2ccd-8456-460b-972c-c9f5d926b5e2) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -8461,7 +8461,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d7cb-6424615322109faf177d642b;64895d45-121c-4009-b15c-675632db00d5) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66722958-37b0a1f4319736795fbaee42;86eb0938-f5f2-4bd8-9b4d-40026ece786f) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -8762,7 +8762,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d631-230dcd9518cf6180310450de;5cf1c183-bdda-44e6-b427-47ec3549dfe1) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667227d0-45a1ea8b1d4a74021cd6cac6;90488a3f-b87e-45cc-9fb5-7c0ba52c77b8) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -8847,7 +8847,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e6d3-40d238e37a7b956a3658228e;07afc1f0-dffa-43e0-b24a-7aa3509b092f) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722ef6-471c4e3a01b5520703a5b75d;1f4bf7d0-28ee-4529-aa79-8ae6868f1d5e) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -8963,7 +8963,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e609-0ea35221113d393025af4796;4b04f458-d2b2-4176-8f88-4873069bfe3d) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722e2d-127ec87d07ba87416675e254;6eaac5fc-26f9-454a-9108-a5dfbfd8aeaf) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9107,7 +9107,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e815-3bdcd0c640944d8d6abbf70f;0f85c3a8-28d4-4a36-b2b5-b118d3dd80d4) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6672303a-13a31fb8052a2e1935d2be00;bce85a12-0195-4316-a036-399e48d9c39e) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9178,7 +9178,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e5e1-3399b02f4c3630bc3377e2f4;f8eabdcd-596e-4277-85c1-1c20a02b4721) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722e06-1a4c598a0ca4244554f151e6;cc53e0b6-b6d2-49ce-a538-90c595e4c99d) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9249,7 +9249,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e682-48b9bd1d0dc23d907558cbe4;4c318d9a-fdd6-4bcf-858e-984244999eda) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722ea7-678799b506fdf8d404c0e5c3;fa8786c0-c40f-47ad-85ed-6d35ab48f788) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9368,7 +9368,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e9c8-5efca3534b5c84ab296728ad;8f35f1a7-f417-4481-9dd1-b21145250aa4) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667231b4-48ee817642523c433589ef95;d91554f1-5f5b-4fc5-82a0-05ab7eb08253) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -9453,7 +9453,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e6aa-6254ea835cf5b9f1073e6690;4ebde97c-e072-4899-82b0-395c78b4112e) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722ecf-0e9282d90000d2cd2d38e588;22499200-4b6d-416c-bf7d-963a1d36b019) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9515,7 +9515,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d798-1dfc43e7409aaa5a399dea5e;89fbd57e-a826-4364-99f5-5e7c0f08ed3d) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66722930-685625fa4c6eaeac700e9add;009763b3-00fb-4041-a5a7-9b3bbc895aa2) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -9642,7 +9642,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e65a-6a89801b7f1346bd64d50933;38ef97ee-5807-452e-9396-c3db0083f808) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722e7e-75dafde907b94d0a01596aa5;4fd0e4f7-d7ec-471e-9544-cb120e5baccc) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9715,7 +9715,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670d81c-7cb013a2595a11281298964d;cfa49739-64a6-4d44-ab79-71904a526655) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667229aa-7d05fdf343b1de6138d653eb;73c79eee-8933-4f29-a80c-5640029a9411) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9788,7 +9788,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e7c4-30e3aba7224c52001243cdf4;20ec4d2e-4a5f-4de6-8ab1-363eb78d4a8d) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722feb-05bb94e170e174117c55d28a;86517e45-01ff-48bd-b881-9e514dbfd7be) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -9916,7 +9916,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e46e-5f95bfb3137bfa91380a856c;756caa0c-0113-45c7-abef-187f3343ac26) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722cbe-33475f9161d47f5a173cd33d;272185d7-ae1a-4f81-9219-f79a5c597511) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10110,7 +10110,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e414-564547120400a78525181c7a;fb1c0bb6-921e-44ec-ac76-8191cbc005da) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722c6c-2a05908827c475423ebb3764;3e85c874-7cea-45ee-9ddd-977ee02cc9c6) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10180,7 +10180,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e8b7-79cb110405bd01690909e0b1;fb881c8d-9c62-4328-9e70-4ddc9f9f6f6f) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667230b3-7e72c58a16a1c46601062ce9;0de11006-f7de-4b38-bf44-121117719094) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10250,7 +10250,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e84d-395f8e807424f414565841db;8bb28ac2-ae64-4459-b91b-562c754a1b6c) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66723062-7b7879ca10a760830e21329f;8f2052ff-12a9-4d4f-88a1-757594a79e0e) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10350,7 +10350,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e72a-1c994828280033ef66d06bee;e6d18fff-c2c4-4816-968d-3aed825c18d9) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722f4e-4ba23d872eea985d4d15a79d;faf8b004-5be5-44a9-8f34-d5c9d2e1a56c) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10421,7 +10421,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e7a3-5b2e63e27aa206d53a733b92;f9dd0a10-b927-478c-bc56-3d19b5d668d7) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722fc8-5132043d2b05fa433d0fd7f7;9de55f19-9dd8-4943-b6e2-5110ca9f69df) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10605,7 +10605,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d7fa-0482150f770ebb4b6e74d178;223b00c6-129b-4af7-b9fd-7e73bb3db200) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66722987-28688485419523000b38c744;7203d1d4-41bd-4e10-8ad2-033b852bdcd4) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -10694,7 +10694,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e752-6edc61576f4f2a535a6c2ba5;059d2da6-1a4a-43c5-b98f-50ad96f3c15d) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722f75-6aee5f2b2f67b2a33562062c;7f6c85bc-cd94-4b4b-9b5f-1f4f6ee16905) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -10754,7 +10754,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d7d2-6bd7057b110d2a6b311f683f;5cc5dc41-f0c5-4787-b66a-67430a31236c) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6672295f-34299c64110921a30a2c124f;a29bcaa5-0064-4c33-a5a8-0dd168574ea3) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -11055,7 +11055,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d638-4434084e7f4fb85d112c4add;22d34d52-e979-4e4a-a6f9-f7eed4ccdd28) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667227d6-5ee67fd54a98fc58786f3438;c3f4df8f-6198-41ab-88b4-58b935402e32) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -11140,7 +11140,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e6d9-6c55cb31181ff756705783ee;7173ed94-3724-4ac5-813b-e266ca3923cd) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722efe-2e29c29a07ff1d52706218bf;b7dabb17-ca1e-40a9-a4a2-43df87328107) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11256,7 +11256,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e610-0bc5abf654a63df43dd35187;84e75c09-b8b7-48ec-8ec1-e4313c2c859b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722e34-14fda9ad65788b9146b892bb;2041b154-10cd-4dfa-921b-373b1745377a) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11400,7 +11400,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e823-2d99ea970aef1cff7f710b90;aa7b0fb2-a68e-4d60-ae6d-fd10739b34b6) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66723041-77f3365d4131648c08ae4284;16b1c3d7-8f62-4651-930a-476d0e96ab12) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11471,7 +11471,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e5e8-1b52fd9d58f8cd06510c08eb;685ff270-0bd6-4030-9473-106ce84c6c3b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722e0c-49f532fe29686c141fdfa836;456e57de-efca-43a7-863d-1c17ddee27d0) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11542,7 +11542,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e688-2bfba3af72305086635af116;b42568a9-645d-4ab3-a0c6-df5293384278) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722eae-550141f9476d94e219d0b20a;c75f342d-f9db-43f7-8a02-8f196ab188fa) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11661,7 +11661,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e9cf-01f023a0275ef861793195a6;0bc50725-bd21-4443-a57c-e11b4a8cc901) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667231bb-658fed82771207c447e17264;5df2ca28-0eea-49f8-9c43-3a83392b31ab) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -11746,7 +11746,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e6b0-03cedae11d37b5673dbab861;2edaf4a2-712f-4169-946e-28ff44a621e1) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722ed5-644c10946ad91c5a7de59072;b5476738-16ac-4fcd-aefa-6c516ba6d83d) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -11808,7 +11808,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d7aa-065b6ae039df0126778fcfcf;067a0fa5-28cb-44e7-8a3e-c30c8ef29229) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66722937-37e615627f56563012796a11;43e3d1d2-1218-4ed7-83b5-41047f767d1a) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -11935,7 +11935,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e661-48e54e1a55db658e15a18c1d;c038020e-17e9-4f7d-9fb1-f35960ee1f90) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722e85-0766406811f16e004da03818;b0e45f6a-8b37-448d-be8f-8c1fdcb1ffe6) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12008,7 +12008,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670d823-2ca90f5f4bd0d03d4515874a;07c49e87-ab5e-4d89-9be7-c641fba88c00) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667229b1-03db939336a37fe26a159c1c;d99c2c28-3214-48a2-a557-bd34910463f0) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12081,7 +12081,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e7cb-5347d5b011b0448f5bc59179;21e18ee0-4699-42eb-948b-977319fab2ac) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722ff2-2577d2de45eff74a2ed37ca2;c92b5c8d-faac-4bd1-aabc-915379b8cd53) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12209,7 +12209,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e478-5008a0d47709cc06244e826e;da46e1ec-b7f4-4915-a457-4f9fc6a81604) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722cc6-7c4fb7bc2f7f728d1b957b49;e2e02bc6-95b0-48c4-9b5c-578f506d56a8) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12403,7 +12403,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e41b-6f9412b7463a988d4fb75140;38a752fe-8d90-467f-bc23-969471e3c647) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722c73-70c42d1f112347580c776d43;43cfa525-b7e5-4497-802b-ab045f4596b1) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12473,7 +12473,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e8bf-596d81a648af4e5350f541ef;8217c7b0-ec92-4de5-a807-7e1097c6d6ca) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667230ba-7b0ba0f11c4b105368f718a7;7cea19ca-a143-4445-8fd2-72e4353e0df3) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12543,7 +12543,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e854-46e6275a1e459ef240a260a8;8247c021-deb4-427a-ad27-eb1100014ceb) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66723069-3b5162d44bcf4fd82dfb5331;980689f6-40eb-4ef7-96b2-87505ce05102) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12643,7 +12643,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e737-78b0544b678676b768acd183;f899e8e1-9922-4b53-8d5e-aa727ad43e0b) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722f5b-587395cd77ad3bbf66d8f7e7;52c28efe-8518-48d9-ab59-98a51c095c2e) Repository Not Found for url: https://huggingface.co/8/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -12742,7 +12742,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e7b0-03453e5d712085645d23dec2;e766bff3-194e-4870-bcf8-c0a05b9af713) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722fd5-0c23651f01cc95924c70f55d;8b946abf-5df6-4257-8cd4-4fc0c0896655) Repository Not Found for url: https://huggingface.co/2/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -13010,7 +13010,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d808-531128e44dbb7754619cc965;0376f505-b035-49c6-a018-eb146468608b) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66722995-23ffa50c26775f2030168773;24acad81-3768-4850-a581-562bfcb925d3) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/recurrentgemma-2b/resolve/main/config.json. @@ -13215,7 +13215,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e75f-63cdeb526ca9cdca3c444a94;f30fc67e-bdaf-4148-9a73-69f4a3cd471d) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722f82-112df0d217bbdb6f786bbe63;e677eeab-8d5e-4b1f-875f-6bedf3f2b0ca) Repository Not Found for url: https://huggingface.co/x/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -13275,7 +13275,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d7df-0552cfd064e4b79025cbccf7;3bcb6f9b-97ea-4f1a-962b-cdfbedff49f4) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6672296d-7c9113515100cdcd6d17aa54;8ad01ecb-79e1-486f-ad2f-2e81c001c02e) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-7b/resolve/main/config.json. @@ -13800,7 +13800,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d646-7621afd86a2b95f83ab325a5;401d3be0-a311-4b8f-8e54-c06cc59c267c) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667227e3-3691407a5886f85638d7f3c3;96d55fe9-be0f-4809-a6b6-85f910549987) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/databricks/dbrx-base/resolve/main/config.json. @@ -13885,7 +13885,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e6e7-4122741b3d6c6f363be61fd2;efe7af33-ef23-45f2-ab45-e72d18806167) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722f0b-4928b30d3b884c75238b7178;be277b07-2ff9-4add-8002-1b0d8d9a57b8) Repository Not Found for url: https://huggingface.co/l/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -14029,7 +14029,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e61d-41adde0d35cdc10261d99ce7;9656a3de-abcf-441e-af33-d4b09ac8fef1) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722e42-6e43b23a5fef0d1c7837a8e3;f1858ec3-9b76-4770-aa28-bc93d674f679) Repository Not Found for url: https://huggingface.co/i/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -14201,7 +14201,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e835-57e10b621862fe912d281c13;69a0facf-a74f-4434-a12a-fc9e2e6b3cfc) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6672304e-68228f380de25bf333d24a85;abe8b82b-c1eb-488d-8614-3bcc2f1a6ccc) Repository Not Found for url: https://huggingface.co/v/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -14300,7 +14300,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e5f5-083264596d0e3ec40622a860;3874cb2b-805c-4c7e-b72e-0567e1804fe0) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722e19-518e58dd40fec6ca30c5a6db;3e6b7ca5-719b-4cb2-ae62-3cc8491e5404) Repository Not Found for url: https://huggingface.co/M/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -14399,7 +14399,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e696-2f5e48224849abd1519e1557;a0adae52-d2ab-4e2b-a4d6-1824ee7bfa70) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722ebb-259a95ea4032f4fb0956dce6;8191fa83-bc74-474d-936e-f2996b675572) Repository Not Found for url: https://huggingface.co/r/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -14490,7 +14490,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670e9dc-1a8a06ac1d199a5c037d126a;0b07d4cb-4c47-4783-8401-154595518474) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-667231c9-1e94cfd63641faa86e408bb8;7389a052-2a45-444e-aa6f-a24ed7330f91) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/tiiuae/falcon-180B/resolve/main/config.json. @@ -14575,7 +14575,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e6be-2d88d6e27045eb320831da3c;29df8040-dd8f-4996-a6f6-83c8293a7735) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722ee3-340211374e0fb28c3004154f;eff40da2-78ce-4157-a458-78ab05b041a4) Repository Not Found for url: https://huggingface.co/a/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -14665,7 +14665,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 367, in hf_raise_for_status raise HfHubHTTPError(message, response=response) from e -huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-6670d7b7-4b55968b0954139166a7ccd3;f15e8e98-b4f7-4fc0-a8fe-645496e4e12c) +huggingface_hub.utils._errors.HfHubHTTPError: (Request ID: Root=1-66722944-7542579d3776fc0c392ec0c7;6734dadc-ef4a-4945-85c2-3459334c6c42) 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository.. Cannot access content at: https://huggingface.co/google/gemma-2b/resolve/main/config.json. @@ -14848,7 +14848,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e66e-5cfd3fe16d0c476a7eadd055;45969fdb-02cd-4098-b806-cad2e5dd0ef6) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722e92-2c73cb1d36e0b374116c439f;510171e7-a160-47bb-a782-9a88992027c6) Repository Not Found for url: https://huggingface.co/t/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -14977,7 +14977,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670d830-622cd6eb66bca9f53ea1d681;3409feb9-9829-4acc-b737-826f86b44261) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667229be-7909317678282174612d3bfe;a4ff448d-d9a9-42a7-923e-0d2aaba23745) Repository Not Found for url: https://huggingface.co/google/recurrentgemma-7b/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -15106,7 +15106,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e7d9-5a2f378f3dbd23eb74e5010a;6968c0be-c00f-4f10-9a7a-91234b5da131) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722fff-54cbe02811452b1402d96ea3;5c394f9b-dd65-4108-93ff-eeccea9bf13f) Repository Not Found for url: https://huggingface.co/B/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -15206,7 +15206,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e487-2913b802411fa49575848a62;893f7ce9-0012-40ed-87ea-5a921d7ec76c) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722cd3-54c0f8d472b2d7d3522ceb4f;135f3525-145f-435f-8811-d48a816179f5) Repository Not Found for url: https://huggingface.co/s/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -15555,7 +15555,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e429-10e69ddb2ecf679c34ff7f2a;a9c4f355-4469-49be-85cc-a444fdb0295d) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66722c81-3dcfdb36698d6f6477c72874;31f75b7f-eb78-4f54-bfa6-5a7b3c05f8b0) Repository Not Found for url: https://huggingface.co/m/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -15625,7 +15625,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e8ce-6b75b9030643d3775c4fb118;cfb334d2-3a06-4dd6-a5da-201d4ad139a0) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-667230c7-717a44bb24c540985ea05d13;2235b792-b08e-446a-8914-60f40a01470c) Repository Not Found for url: https://huggingface.co/1/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`. @@ -15695,7 +15695,7 @@ Traceback (most recent call last): hf_raise_for_status(response) File ""/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_errors.py"", line 352, in hf_raise_for_status raise RepositoryNotFoundError(message, response) from e -huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-6670e86b-1ba175c27622547e7edc0e86;1e0e9503-aaa5-4f1a-970a-dffa40e3a7bf) +huggingface_hub.utils._errors.RepositoryNotFoundError: 404 Client Error. (Request ID: Root=1-66723077-16f8f18e78e1014e77a90a23;ff8b1a09-eed7-4865-b98d-f1bff2076774) Repository Not Found for url: https://huggingface.co/0/resolve/main/config.json. Please make sure you specified the correct `repo_id` and `repo_type`.